Ruchir Garg

Introduction Education Experience Publications Teaching Skills Achievements

prof_pic.png

To love is to suffer and there can be no love otherwise.

Hello! My name is Ruchir Garg, and I absolutely love to solve puzzles, play racing games and philosophy admirer and deeply inspired by Fyodor Dostoyevsky .

I am currently working at Neuravest Research. as a Quantitative Analyst, primarily focussed on developing Thematic multi factor equity strategies and intraday futures strategies.

I have developed several strategies by extracting alpha from different datasets to create trading signals, some of the datasets being

  • Institutional Flow data
  • Order Book Data
  • Brokerage inflow/ Outflow and DarkPool Data
  • Sentiment/ News and Analyst consensus Data
  • Fundamentals (visible alpha, earnings, statements)

In addition to my interest in finance and trading, I've cultivated a deep appreciation for other areas of study, including:

  • Genomics (transcriptomics, proteomics):
    • My research experience at UCSD-Health during my master's program involved working on genome assembly for both long and short noisy reads.
    • At the La Jolla Institute of Immunology, I engaged in immunological research, focusing specifically on TCR-Antigen Affinity.
  • Low Latency Software Optimizations: My tenure as a Senior Software Engineer at D.E.Shaw & Co was an enriching phase where I developed a keen interest in low-latency software engineering. This role allowed me to explore and solve diverse challenges, particularly in areas like asynchronous processing, cache management, and data structure optimization.
  • Quantum Computing algorithm design: I am eagerly looking forward to opportunities to engage in this evolving field.

If you have an idea and want to see if I’d be interested in collaberation / helping out, please contact me (ruchirgarg5@gmail.com)!

Bachelors

Indian Institute of Technology Indore (IIT Indore)

B-Tech Computer Science (June 2018) CGPA: 8.8/10.0

Minor: Econometrics, Macro/Micro Economics

Relevant Courses: Operating Systems, Database Management, Convex Optimization, Computer Architechture, Design & Analysis of Algo., Deep Learning, Discrete Mathematics, Compiler Design

Masters

University of California San Diego (UCSD)

Masters Computer Science (June 2023) CGPA: 3.8/4.0

Minor (Rady School of Business UCSD): Asset Management, Data Science for Finance

Relevant Courses: Prob Stats for Data Science, Advanced NLP, Statistical NLP, Probabilistic Learning Methods, Unsupervised and Semi-supervised Learning, Optimization techniques & Reinforcement Learning, Computational Bio Informatics

Internships

Machine Learning Research Intern

La Jolla Institute of Immunology - Jan 2023 - April 2023
  • Project: Recognition of T Cells Causing Asthma Using Machine Learning Techniques.
    • Advisor: Dr. Vijayanandan Pandurangan.
    • Objective: To identify different T cells involved in Asthma through advanced ML techniques.
    • Methodology: Employed encoding techniques for the structure of T-cell and antigen and utilized an attention mechanism to analyze the interactions between T-cell and antigen.
    • GitHub Repository: ATTENTION_TCR
    • Novelty and Impact: This approach significantly enhanced the identification process of potential T cells compared to existing methods. Using this novel methodology, the number of T cells that can be tested for the VDJb database was substantially increased, demonstrating a considerable advancement over traditional techniques. The findings are currently undergoing wet lab validation to verify the new TCR-antigen binding groups.

Machine Learning Research Intern

Alation - Jun 2022 - Sep 2022
  • Primary Role: Design and build a Federated Learning infrastructure to enhance data access and cataloging.
  • Key Contribution: Developed a secure framework for asynchronous model distribution to multiple clients, focusing on privacy and data security.
  • Algorithm Implementation: Implemented various robust aggregation algorithms within the framework, including FedAvg, Scaffolding, and Secure Aggregation, to ensure effective and secure model training.
  • Client Selection Strategy: Devised a methodology for selecting clients during model aggregation based on the number of learning cycles per client, to maintain consistency with the global minima.
  • Integration with AWS SageMaker: Streamlined the entire federated learning process by integrating it with AWS SageMaker, enhancing operational efficiency.
  • Research Foundation: My work aligns with the pioneering concepts in Federated Learning, as outlined in Google's research blog. For more details, see Google Research Blog on Federated Learning.

Machine Learning Research Intern

UCSD-Health - Jan 2022 - June 2022

Desingn and Build Federated Learning Infra

  • Project Focus: Efficient Identification of Sequencing Errors to Improve Haplotype Genome Assembly.
  • Advisor: Dr. Vikas Bansal.
  • Research Approach: Developed probability models leveraging neighboring reads to accurately process false variant calls.
  • Methods:
    • Utilized Convolutional Neural Networks (CNN) with layers representing different genomic features to enhance accuracy in identifying in-dels in genome assembly.
    • Applied Maximum Likelihood Estimation (MLE) and Expectation-Maximization (EM) algorithms, combined with beam search, to pinpoint false variant positions and reconstruct the original sequence from noisy aligned fragments.
  • Novelty and Impact: This innovative approach challenges conventional methods by exploiting the inherent similarities in human genome structures.
  • Outcomes: Successfully surpassed the State of the Art (SOTA) with an accuracy of 99.7%, significantly advancing the field of genomic assembly.
  • GitHub Repository: For detailed insights, visit HAPCUT_preprocessing.

Research Intern

Max Planck Institute (Magdeburg, Germany) - May 2017 - July 2017
  • Project Focus: Implementation and Testing of Model Order Reduction Techniques on GPU.
  • Advisors: Dr. Jens Saak (MPI), Dr. Kapil Ahuja (IIT Indore).
  • Objective: To implement and validate the viability of various Model Order Reduction (IRKA) techniques for dimensionality reduction, specifically on GPU platforms.
  • Methodology: Implemented algorithms in CUDA, focusing on Block QR decomposition, to optimize and test on several Linear Dynamical Systems such as HEAT BEAM models for assessing correctness and stability.
  • Performance: Achieved approximately 15x amortized efficiency improvement compared to CPU-based implementations.
  • IRKA Reference: For more information on IRKA, visit Wikipedia.
  • Algorithm reference: Algorithms implemented can be found on the MPI-Magdeburg MORLab Project.
  • Research Report: Detailed insights are available in the research report, accessible here.
  • Conference Poster: The conference poster showcasing this research can be viewed here.

Engineering Intern

GeeksforGeeks - Dec 2015 - Jan 2016 New Delhi Area, India
  • Key Project: Optimization and Enhancement of the Internal Search Engine.
  • Technologies Used: Elasticsearch.
  • Improvements Implemented:
    • Engineered solutions for handling queries with typos and incomplete words, making the search engine more robust and user-friendly.
    • Developed case insensitivity in search queries, enhancing the search functionality for a wider range of user inputs.
    • Optimized search algorithms to return relevant results based on user ratings, page views, and user feedback, significantly improving the accuracy and relevance of search results.
  • Impact on Customer Reach: These enhancements attributed to a substantial increase in customer reach and engagement, as reflected in improved user satisfaction and higher interaction rates with search results.
  • Contribution to Topic Modelling: Implemented topic modelling to group related articles, which streamlined the content discovery process for users and aided in efficient information navigation.
  • Company Growth: Interned during the nascent stages of GeeksforGeeks, which is now recognized as one of the largest repositories for technical blogs globally.

Full Time

Quantitative Analyst

Neuravest Research - Jun 2023 - Present
  • Intraday Long-Short Equity Strategy:

    • Objective: Developed an equity-based intraday long-short strategy using institutional flow data.
    • Performance: Achieved a live Sharpe ratio of 3.7.
    • Key Technique: Leveraged order book data for strategic exits.
  • Beta Neutral Sentiment Strategy:

    • Data Source: Utilized DowJonesNewsWire for sentiment analysis.
    • Approach: Combined sentiment signals from an instruct LLM model with technical signals using convex optimizaiton, and used Markowitz optimization for allocations.
    • Hedging: Employed sector ETF shorts for hedging long positions.
    • Performance: Achieved a live Sharpe ratio of 1.93.
  • Event-Based Sentiment Strategy:

    • Feature: Developed sentiment features based on text and topic analysis.
    • Hedging: Strategy hedged with sector ETFs.
    • Control: Adjusted signal combination coefficients based on market regime and fundamental features.
    • Performance: Reached a Sharpe ratio of 2.33 in testing phase.
  • Current Project: Working on a reinforcement learning-based allocation algorithm for the sentiment strategy.

Sr. Software Engineer (Python/ C++ Infrastructure)

D.E.Shaw & Co. - Jan 2020 - Jul 2021
  • Developed advanced sorting algorithms for multi-column data in Cython C++, accommodating custom objects like NAT (Not a Date), achieving a 5x amortized efficiency gain.
  • Redesigned the strategy testing overlay system from file-based to key-based, leading to a 10x improvement in data freshness and robustness:
    • Replaced initial file replication system to reduce stale data issues.
    • Implemented a lazy update mechanism for cached data in strategy classes, keyed to specific requirements.
    • Introduced diverse file-based systems (H5, Pkl, Redis backends) within the overlay for optimized storage and readability.
  • Designed and maintained the Portfolio Optimization Pipeline Implemented Proximal Policy Optimization based Reinforcement Learning strategy.
  • Executed optimizations for the storage and querying of extensive historical data (up to ten years), resulting in a 6-7x improvement in data retrieval times:
    • Applied sharding on custom key axes and parallel chunked updates.
    • Research and deploy repartitioning schedule of h5 for non-contiguous data, realizing (1.5x amortized efficiency)
    • Refined data alignment with non-prefix keys using almost-dense vs. sparse optimizations during cache retrieval.
  • Streamlined data ingestion and preprocessing pipelines; assisted in establishing partnerships with alternative data providers.
  • Software Engineer (Python/ C++ Infrastructure)

    D.E.Shaw & Co. - Jul 2018 - Dec 2019
    • Engineered a library for the serialization of 'exotic' Python objects for distributed computing, achieving a 100x performance advantage over solutions like dill and cloudpickle.
    • Automated the sharing of private data linked to strategy classes, enhancing team collaboration and efficiency while maintaining code confidentiality.
    • Made significant modifications to OpenMP, MPI, MKL, HSL, and IPOPT libraries for uint8, uint16 object support, resulting in a 4x improvement in benchmark efficiency.

    Publications

    A Fast Adaptive Active Learning Using Kernel Ridge Regression and Clustering for Non-stationary Environment.

    Read the Paper

    Online Meta-Cognitive Type 2 Fuzzy Kernel based One Class Classifier for Non-stationary Environment

    Read the Paper

    Teaching Experience

    DSE-220 (UCSD Professional Degree): Data Science for ML

    Teaching Assistant

    Spring 2023
    • Teach experienced professionals with the advances in ML and data science.
    • Created assignments and held weekly office hours to discuss research projects and doubts.

    CSE-202 (UCSD Graduate Course): Algorithm Design

    Teaching Assistant

    Fall 2022
    • Held problem solving sessions (twice per week) to tackle new algorithm problems.
    • Held weekly office hourse to discuss assignments and course doubts.

    CS-507 (IIT Indore Undergraduate Course): Software Design

    Teaching Assistant

    Spring 2018
    • Held weekly office hourse to discuss assignments and course doubts.
    • Create assignments and aid students with different Software engineering design tools.

    Programming Club (IIT Indore)

    Algorithms Coach

    Spring 2016 - Spring 2018
    • Coaching aspiring competitive programmers for ACM-ICPC competitions.

    Skills

    Languages and Databases

    Python, C++, CUDA, MATLAB

    Tools/Technologies

    Git, LaTeX

    Frameworks and Libraries

    Pandas, PyTorch, NumPy, SciPy, Scikit-learn, Pandas

    Scholarships and Achievements

    Research Grant

    Jan 2022 - Jan 2023

    Received a research grant from the National Institutes of Health (NIH). Additional information is available on the NIH page.

    PRIUS Fellowship

    Jan 2018

    A prestigious fellowship under the program Promotion of Research and Innovation for Undergraduate Students. More details can be found on the PRIUS page.

    KVPY Fellowship

    2013, 2014

    Awarded the Kishore Vaigyanik Protsahan Yojana (KVPY) fellowship, an initiative of the Department of Science and Technology, Government of India. For more information, visit the KVPY Page.

    Aryabhata Mathematics Competition

    Silver (2008), Bronze (2010)

    Achieved Silver and Bronze medals in the Aryabhata Mathematics Olympiad. Details about the competition can be found on the Aryabhata Maths Olympiad Page.

    ACM-ICPC

    Rank 14 (2017 regionals), Rank 63 (2016 regionals)

    Secured high ranks in the regional rounds of the ACM International Collegiate Programming Contest (ICPC). Learn more on the ICPC page.

    news

    Dec 1, 2023 Website Created!