Second, when handling complex input data with intrinsic structure, I encoded the input to a latent space. Then, the encoded input was transformed with respect to the intrinsic structure. These ideas allowed the algorithm to handle the complex input much more efficiently and observe the 3D robotic-arm’s and 3D protein’s structural properties when the input was a 2D image or a 1D protein sequence.
How do you optimize a parameter that has no derivative? Does an optimum even exist? I encountered this key question many times while finding: optimal model architecture for predicting satellite images, best loss function for protein structure prediction, effective estimations of cost-to-go functions.
I know that the program will be a challenging and require me to grow beyond my comfort zone. My experiences of overcoming difficulties in academics, industry, and self-study makes me well-prepared.
Completing the master's degree in Applied Mathematics and Computer Science gave me experience of applying logical and theoretical concepts. Courses such as Linear Algebra, Algebra, Analysis, Operations Research provide a solid background to understand AI ideas at a fundamental level. Analyzing protein structure using CNNs and persistent homology in my thesis, gave me interdisciplinary insight on how math and CS complement each other. Finally, I really enjoyed the many hours thinking, meditating, and exploring theoretical concepts and questions.
Working in the industry gave me valuable experience overcoming difficult problems and producing consistent results. Here are three examples.
Pursuing challenging independent research interests during my self-studies taught me discipline, motivation and persistence. Studying the AlphaGo Zero paper, I implemented a Monte Carlo Search Tree model-based RL to play variations of tic-tac-toe with longer length-to-win and boards. It was very challenging to get the initial prototypes to work on larger variations of the game; I carefully and patiently tried many different ideas. Eventually, reasonable success was found using: multi-step lookahead using model value estimation, augmentation of board inputs using rigid transformations, adjustments on rollout sample size depending on game depth, and many minor optimizations. Wanting to understand the RL the much better, I am currently studying the fundamentals of markov decision process programming and optimal control.