Touch Typing

The one, simple thing you can do to increase your productivity more than 10-fold is to learn how to touch type. 

This isn’t just about improving your performance at the keyboard. It has to do with removing distractions and obstacles when coding. It’s about becoming one with the machine, translating the thoughts in your head into code, without having to hunt around for the right key and pressing it one at a time. Your brain works much faster than you can express your thoughts, and anything you can do to bridge that gap improves your performance. 

And this has a multiplier effect as you learn how to use a good editor.


Follow by Email

The Developers Bookshelf

Algorithms & Data Structures

I’ve found multiple perspectives deepens my understanding of topics. The following books cover roughly the same ground, but have very different approaches, and I’ve gotten different insights from each of them

  • Introduction to Algorithms, Cormen, Leiserson, Rivest and Stein. The basics, written clearly and yet with rigor, this book should be the one to start with on most topics. Read it for self-study or as a reference.
  • Algorithm Design, Kleinberg and Tardos. A much more detailed look than CLRS, and the systematic exposition is a joy to read. There are a crazy number of exercises as well!
  • Algorithms, Dasgupta, Papadimitriou and Vazirani. Much less detailed than the first two, but a very different approach to many topics. 
  • Algorithms, 4th Ed., Sedgewick and Wayne. Many fewer topics covered, but in a lot more depth. Especially see how to analyze and test implementations. Great exercises and see the associated booksite as well.

And of course, no list of Algorithms books can be complete without

  • The Art of Computer Programming, vols. 1-n, Knuth. This is advanced, and not for the faint of heart. Also, the best categorization of sorting algorithms that I’ve seen. And demonstrates the limits of analysis (by which I mean it demonstrates how much analysis you can do!). 

Programming Praxis

More than syntax, programming is about the mindset. Here are a few books that help in this area:

  • Programming Pearls, Bentley. Yes, it’s old. Yes, it’s worth every bit.
  • Beautiful Code, Oram & Wilson. Bringing art into engineering
  • The Pragmatic Programmer: From Journeyman to Master, Hunt & Thomas


  • Hackers and Painters: Big Ideas from the Computer Age, Paul Graham
  • Turing’s Cathedral: The Origins of the Digital Universe, George Dyson. A bit of history. 
  • Anything and everything by Simon Singh
  • “Surely You’re Joking, Mr. Feynman”, Richard Feynman

Language-Specific Books


  • The C++ Programming Language, Stroustrup
  • Accelerated C++, Koenig & Moo
  • Modern C++ Design, Alexandrescu


There are too many (very) good options, but I’m going to stick with just one:

  • Python Data Science Handbook, Jake VanderPlas


  •  Design Patterns: Elements of Reusable Object-Oriented Software, Gamma, Helm, Johnson, Vlissides
  • Head First Design Patterns, Freeman & Robson


  • The Hitchhiker’s Guide to the Galaxy, Douglas Adams. Everyone should know the answer to the ultimate question
  • The Discworld novels, Terry Pratchett
Follow by Email

Round Ireland with a Fridge, Tony Hawks

A cheerful account of a pointless trip that brought to mind something that Jerome K. Jerome would write. I kept reading the dialogues in Derry Girls accents. And there are gems like these:

‘Yes’, I replied accurately.

Follow by Email

Programming for the Programmer


Students aspiring to program come in all varieties. Some are good, some are not so good and this categorization is strictly within the same pool.

The frustrating part (for me) is that the good programmers have a long way to go to become competent developers. And the not-so-good programmers are not so good for no fault of their own. I find that the assignments that they do, either in college or elsewhere are unambitious and are too focused on specific concepts.

Many students seem to think sites such as leetcode are all that is required. I beg to differ – these are fine for preparing for interviews, but not for  becoming good developers.

Here’s a series of assignments that will take you beyond your current capabilities will teach you more about algorithms and data structures than you currently know, help you learn how to use the right tools, techniques, and skills, and get you started on building your portfolio of work.

The first group doesn’t even ask you to write any code:

The next group is focused on the basics:

  • Sort
  • Search
  • Fractals
  • Command line / options

And a few complicated ones:

  • Shell
  • Bellman-Ford
  • Matrix multiplication
  • Random walks

And the skills needed for tomorrow today:

  • Shared-memory multiprocessing
  • Distributed-memory multiprocessing
  • Map-reduce
Follow by Email


What you will learn:

  • Software: practice with loops and operations on arrays
  • Domain: approximations and error measurements, sampling

Input: Construct a (continuous) signal of your choice – say a simple sine
wave with a particular frequency and amplitude. Plot this. Note that since you
are representing the continous signal on a computer, so you are already
sampling it.

Discretization in space:

  • For the amplitude, choose a $n$ and divide the range into $2^n$ steps. For each index of your signal, (i) round up (ii) round down and (iii) round to the nearest step value.
  • Calculate the mean squared error due to this process.
  • Repeat for different values of $n$
  • Plot the error for different values of $n$. Does this match what you think it should look like?

Discretization in time:

  • What happens if you take fewer / more samples in time for representing your signal? How would you now determine the quality of your representation?
  • What is the lowest you can go? How does this relate to the frequency of the signal? (*cough* nyquist *cough*)

Discretization in both dimensions

  • Try it out!
Follow by Email


What you will learn

  • Software: relatively simple constructs, but a great exercise
    in constructing iterations. Visualization.
  • Domain: develop an intuition for how convolution works.
    And its time complexity

Input: construct the following signals of roughly the same height and width. Consider if each of these are odd or even signals.

  • Single rectangular pulse centered at zero
  • Single triangle centered at zero
  • One period of a square wave
  • Single sawtooth
  • Make your own!

Convolve each pair of inputs

  • Decide what the range of the summation should be
    ($-\infty$ to $\infty$?). i.e., how many steps do you need?
  • Start with a single step and plot the result. Repeat for each step. Try animating it!
  • How does the type of signal affect the output?
Follow by Email


What you will learn

  • Software: fairly simple constructs
  • Domain: Modulation, FSK, tradeoffs

Amplitude Modulation

  • Construct a carrier (high frequency) and a signal, represented as a list of values
  • Simply add the two
  • Visualize all the three signals
  • Take the modulated signal and subtract out the carrier to
    recover the original signal

Frequency Modulation – 1

  • Let’s start with a binary data signal.
  • Based on the frequency of your carrier, determine
    an appropriate $ \Delta_f $
  • Modify the carrier frequency for a range of cycles based on
    the input signal
  • Plot all three signals
  • Recover the input signal

Frequency Modulation – 2

  • Rather than a $f_c \pm \Delta_f$ for a $1$ and $0$, use a return-to-zero scheme. So the frequency changes only if there is a change
    from zero to one or vice-versa.
  • Plot all three signals
  • Recover the input signal

Frequency Modulation – 3

  • You could group your input data by two bits (so four levels) and use correspondingly more $\Delta_f$
  • Plot all three signals
  • Recover the input signal
Follow by Email

The Rich Man’s Square Wave Generator

What you will learn

  • Software: python, matplotlib
  • Domain: a different way of looking at Fourier series

Do this in python, mainly for the visualization capabilities

Rich man’s square wave generator

  • Create a sine wave of a given frequency and amplitude. Plot it for one period
  • Create a second sine wave of twice the frequency and half the amplitude as the first. Add the two cycles to the one from the previous step.
  • Continue with a third, fourth, … nth sine wave of successively higher frequencies and smaller amplitudes.
  • What waveform do we get as we add higher frequency components?


  • Plot the error between your square wave and a true square wave at each step. Can you bring this down to zero? what is the limit? Can you prove this limit?

Other combinations

  • Take odd multiples of frequency and divide the amplitudes by the multiple squared
  • Take odd multiples of frequency and divide the amplitudes by the multiple squared
Follow by Email

The Poor Man’s Square Wave Generator

What you will learn

  • Software: encapsulating functionality, parameterizing functions
  • Domain: frequency and period of a square wave

The poor man’s square wave generator

  • Decide a frequency and amplitude of your desired square wave
  • Create a list with low and high values (these correspond to the amplitude) and specific indices (these correspond to the frequency)
  • Plot this!
  • Create additional signals of multiples of this frequency, and different amplitudes
  • Repeat using numpy arrays. How might you efficiently create these signals, using features of the numpy library?
Follow by Email

Digital Logic Simulator

What you will learn

  • Software: Depending on how you architect this, fairly sophisticated! Work with graphs, and best done using OOD.
  • Domain: Digital logic circuits, and the sheer exponentiality of 2n2n

This is pretty advanced but will be pretty cool. We will build a digital circuit simulator in this exercise. We will also parallelize it using a pretty cool technique and will add timing analysis and fault simulation as well.

You could do this in python, but if you want to do the parallel simulator, you will need to use C or C++ (maybe Java?)

SimSim (Simple Simulator)

We will create a graph to represent the circuit in this part. To start with, we will create nodes and edges connecting nodes. The nodes will have a basic type (NOT, AND, OR, etc.). The nodes will be connected to each other to form a circuit; each node will have a specific number of inputs but can drive multiple other gates. We will have special nodes called Primary Inputs (PIs) and Primary Outputs (POs).

Each node ‘reads’ its inputs — the values on its input edges and determines its output value based on its type. It places this value on its output edge(s).

Create a circuit, apply patterns to the PIs and see if the POs match what you expect. You should be able to cycle through the truth table of your circuit to verify the output.

FaultSim (Fault Simulator)

A simple fault model is a ‘Stuck-at-‘ model. Tie an edge to a constant 00 or 11 and see if you can detect this at the POs.

Rather than simulating all possible input values, can you figure out a minimal set that can detect all possible faults?

TimeSim (Timing Simulator)

SimSim calculates the output values directly, it has no notion of gate and propagation delays.

Assign delays to each gate (and what the heck, to each connection as well). For example, an inverter can have a delay of  1 time units, a NAND gate of  2 and a NOR gate of 3 time units (hmm, why these three values?). Now, set up a global clock, and calculate the outputs at each time step. Compared to SimSim, the final value of POs will settle after a few time units.

And they may bounce around before settling! Can you set up a circuit so that you see glitches at the output?

ParSim (Parallel Simulator)

This is a unique form of parallelism!

In all of the above simulators, the values are 1 bit at each PI, gate output, and PO. The function evaluation at each gate is also a bitwise operation. However, we could instead treat each of these as a 32 (or longer!) bit int and do the same logical operations on the entire word! This will take exactly the same amount of time as before, but we’ll be running 32 tests in parallel. Do you see what I meant by calling it a unique form of parallelism?


  • Read in a circuit from a file
  • Read in inputs and outputs from a file
Follow by Email

RC Circuit Analysis

What you will learn

  • Software: relatively basic constructs. Iterative solver!
  • Domain: A better feel for a capacitor charging and discharging and the time constant.

RC Circuits

Take as input values of R (in Ohms), C (in Farads), and supply voltage (in Volts). Assume that the voltage is applied to a series RC circuits at time t=0t=0, and calculate the voltage across the capacitor at different times. Plot these!


  • Determine how much time it will take to charge the capacitor to a target voltage.
  • Vary R to see how the curve shifts. Determine the value of C so that the shift is exactly the same. Which has a larger effect?
  • Repeat for discharging the capacitor, rather than charging it


Apply a square wave as input and simulate for a few cycles. Vary the period and duty cycle and see if the behavior matches your intuition

Follow by Email

CMOS Characterization

What you will learn:

  • Software: relatively simple implementation, plus some visualization
  • Domain: improve your understanding of how a transistor works

CMOS Characterization – 1

A CMOS transistor operates in three different regions — cutoff, saturation, and linear. The equations for these are fairly straightforward. Implement these using hard-coded values for parameters that you may need to specify, such as doping concentrations, mobility, etc.

Given input voltages, determine the region of operation and use the appropriate equation to calculate $I_{DS}$ 

CMOS Characterization – 2

  • Calculate $I_{DS}$ for different values of $V_{GS}$ and $V_{DS}$ 
  • Plot $I_{DS}$ v. $V_{DS}$ for different values of  $V_{GS}$

CMOS Characterization – 3

  • Analyze the differences between N- and P-MOS transistors
  • What is the impact of transistor dimensions?
  • Read transistor parameters from a file and repeat for different technology nodes.
Follow by Email

Programming for the Electronics Engineering Student


I’ve had a few odd discussions with students from the Electronics branch related to software development. Broadly,

  • sniff I’m in electronics, don’t expect me to dabble in ugh software
  • We don’t get the subjects that CS students do, so aren’t as prepared for interviews as they are
  • eh, I’m just going to do it

There’s so much to unpack here…

Firstly, software (and data structures, and algorithms, and …) is not the domain of any one branch. It is a tool to be used just like any other and a very powerful tool at that. So if you’re being snobbish about not doing any software at the altar of “electronics,” you’re just shooting yourself in the foot. Most electronics is software, and the more comfortable you are with this tool, the better of an electronics engineer you will be. Plus, software can help you understand electronics much more than you otherwise would.

Second, there are a lot of opportunities to develop software skills. The courses that you don’t get to do aren’t all that important, you can learn the core concepts by yourself. There are a few (Databases!) that you won’t get exposure to, but you haven’t really missed much. Read on!

Finally, self-study is the best way to develop these skills. Here’s the big secret:

the only way to learn how to program …. is to program

The obvious question is: “what do I program?”

This is a series of posts that presents program suggestions, what to look for and what to focus on. If you go through these, you will be a better electronics engineer and a better software developer than 99% of your peers, irrespective of their major. I’ve gone through the typical subjects from the second year onwards, and defined assignments that (i) build on the theory that you learn (ii) expose you to different implementation concepts (iii) improve your understanding of data structures and algorithms

One last thing: the discipline of programming

  • All your implementation should have a reasonable testing strategy in place
  • All your code should be instrumented to measure the performance of key parts of the code
  • All C and C++ code should be compiled with -Wall
  • All code should be valgrind-certified error-free
  • All code should be in a version control system

I’ll talk about each of these points in a later post.

now let’s

Shut Up and Code!

Program ideas on:

Follow by Email

Cibola Burn, James S. A. Corey

Basic problem solving. If you don’t have the data you need, play with the data you have, see if something comes out of it. She’d made it through three semesters of combinatorics that way. All right.

Forwarded to all my combinatorics students 🙂

Follow by Email

Oathbringer, Brandon Sanderson

The author is pretty well known for his epic series, so I decided to give it a try, with the Stormlight Archive. I still have one more to go, but  I don’t think I’ll continue — there were flashes of brilliance, especially at the climax, but the rest was too…. boring.

Especially given that I was also reading LeGuin concurrently!

Follow by Email

  • Search

  • Categories

  • Post Date