Shuo Yin (尹硕)

Quantitative Researcher

Email: yinelon@gmail.com

Phone: (+86) 15690129730

Web: https://yinshuo-thu.github.io

About Me

I am currently pursuing a Master’s in Finance at Tsinghua University, after earning my Bachelor’s in Engineering from Shanghai Jiao Tong University. My undergraduate studies spanned computer vision, NLP, and time series analysis, which now influence my work in quantitative modeling and factor discovery using deep learning.

I am open to opportunities in Quantitative Research and welcome connections for collaboration or discussion. Feel free to reach out via WeChat (yins156) or email at yinelon@gmail.com. For more details, you can download my bilingual resume here.

Education

Tsinghua UniversityTsinghua University, School of Economics and Management

Master of Finance

Beijing • Sept 2025 - Jun 2027

Current Courses: Large Language Models and Generative AI, Financial Derivatives, Financial Data Analysis, Introduction to FinTech

Shanghai Jiao Tong UniversityShanghai Jiao Tong University, School of Biomedical Engineering

Bachelor of Engineering (Electronic & Computer Eng. Track)

Shanghai • Sept 2021 - Jun 2025

Awards: Outstanding Graduate, Merit Student Award, First Class Academic Scholarship

Core Courses: Data Structure (A), Methods in Mathematical Physics (A), Biostatistics (A-), Medical Robot Control (A)

Teaching Assistant: NYU ShanghaiNYU Shanghai - Recommender Systems(CSCI-SHU 381), Machine Learning(CSCI-SHU 360)

Internship Experience

Minghong InvestmentMinghong Investment

Quantitative Research Intern

Shanghai • Dec 2025 - Present

Crypto HFT

Conducted research on crypto machine learning, market-making and arbitrage strategies in the overseas HFT prop trading team

  • End-to-End Return Forecasting Pipeline: Developed an end-to-end return forecasting pipeline for highly liquid crypto perpetual futures by converting L2 order book data into time-series inputs; leveraged multi-scale convolutional feature extraction and a regression head for prediction

  • Model Architecture Optimization: Optimized the model architecture to improve training stability and inference efficiency; extended to a multi-task setup and used attention to propagate short-term signals into longer-horizon forecasts, supporting real-time monitoring for market-making

Lingjun InvestmentLingjun Investment

Quantitative Research Intern

Beijing • Aug 2025 - Nov 2025

Stock Team

Worked primarily on model research and strategy development for A-share equities in the stock team

  • Machine Learning Cross-Sectional Forecasting: Leveraged a 1,000+ alpha-factor library for China A-share equities; built tree-based and deep learning models with rolling training around key trading dates, and delivered >10% improvement over the baseline. The best model achieved 40%+ annualized long-only return in backtests and was deployed to live trading

  • Deep Learning Time-Series Strategies: Built end-to-end time-series pipelines, implemented multiple baselines and a stacking ensemble; focused on architecture optimization, loss function design, and attention mechanisms, with hyperparameter tuning and model combination achieving 50%+ annualized return in backtests. Several baseline models were deployed to live trading

Ant GroupAnt Group (Alipay)

LLM Algorithm Intern

Shanghai • Apr 2025 - Aug 2025

Ling 2.0 Technical Report

Engineered ProQ model for automated data screening on Ling LLM; fine-tuned quality-filtering LLMs and streamlined the pipeline

  • Data Screening Model: Built a custom agent-based data screening and harmonization model (ProQ), pre-screening data using metrics like PPL and iteratively selecting representative data for fine-tuning quality-filtering models. Post-trained Ling MOE Lite on ~3M optimized data, improving language understanding and code completion by over 1%
  • Signal-Based Data Filters: Developed rule- and model-based operators to identify low-quality data using targeted signals

ProQ Framework

Panoramic Hills CapitalPanoramic Hills Capital (AUM $4B Hedge Fund)

Crypto Research Intern

Shanghai • Jan 2025 - Mar 2025

Conducted crypto market monitoring and developed sentiment indices on crypto, providing insights for fund positioning

  • NLP Sentiment Index Modeling: Engineered quantitative sentiment indicator; curated datasets from X/Crypto News, LoRA-fine-tuned FinBERT on 8k labeled samples (cross-entropy loss, F1=0.91), yielding normalized daily scores [-1,1] to guide BTC trading
  • Crypto Analysis: Monitored daily crypto trends; constructed financial models for BTC miners (MARA, CLSK, RIOT, etc.)

Research & Publications

TS-Agent: RL Empowered LLM Agents for Financial Time Series Forecasting

GitHub Repository

Jun 2025 - Sept 2025, ICASSP 2026 In Submission

Authors: Shuo Yin, Jiahong Zhu, Tianlong Yang, Yuqiao Liu

Developed TS-Agent, a reinforcement learning–based agent for financial time-series forecasting, using LLM-generated strategy pool for exploration-enhanced fine-tuning, then seeding stepwise RL from selected strategies with custom RL module.

TS-Agent Framework

QuantEval: A Benchmark for Financial Quantitative Tasks in Large Language Models

Paper

Sep 2025 - Present, ACL 2026 In Submission

Authors: Zhaolu Kang, Junhao Gong, Wenqing Hu, Shuo Yin, Zhicheng Fang, Yingjie He, Rong Fu, Eric Hanchen Jiang, Leqi Zheng, Xi Yang, Xiaoyou Chen, Richeng Xuan

QuantEval is a new benchmark designed to evaluate Large Language Models (LLMs) in financial quantitative tasks. It assesses three areas: knowledge-based question answering, quantitative reasoning, and quantitative strategy coding. Tests on many top LLMs show major performance gaps. To address this, we apply supervised fine-tuning and reinforcement learning to improve results.

QuantEval Framework

AlphaMaster: Comprehensive Alpha Discovery and Agent Implementation

GitHub Repository

Aug 2025 - Present, KBS In Submission

Integrated Agents, Genetic Algorithms, and Flow Networks to produce diversified alpha factors Developed factor alignment and filtering framework for A-shares and cryptocurrencies, implementing data cleansing, dual-chain dynamic iteration, and agent-based reporting

EVR: Eyesight-Only Visual Representation for Recognition without Language

Paper

Aug 2025 - Present, CVPR 2026 In Submission

Authors: Rong Fu, Chunlei Meng, Shuo Yin, Zijian Zhang, Weizhi Tang, Shaobo Wang, Youjin Wang, Simon James Fong

This research presents a vision-based approach that abandons text and allows visual features to communicate through a graph of image prototypes. Using self-supervised learning, the model uncovers emergent semantic structures, where meaning is determined purely by vision. Text is only used for interpretation, not during training. This method demonstrates that vision alone can recognize and name the world without relying on textual input.

EVR Framework

Revolutionizing Open Set Domain Generalization: Learning to generalize via Conditional Diffusion Without Gradient

Paper

Jul 2025 - Present, CVPR 2026 In Submission

Authors: Haidong Kang, Shuo Yin, Huiquan Zhang

This work introduces OGDiff, a novel OSDG framework that replaces expensive second-order meta-learning with a conditional diffusion process. This avoids inner-loop backpropagation, reducing computational costs and improving stability. The MSDFF module is also introduced to better capture both fine- and coarse-grained features, enhancing recognition of known and unknown classes.

OGDiff Framework

LNG Freight Rate Multifractal Analysis and Trading Strategy

Published Paper

Feb 2024 - Sept 2024, Published in Nonlinear Dynamics (SCI Q2)

Authors: Feier Chen, Shuo Yin, Jiahang Zhang, Yi Sha & Huaxiao Ji

Based on MF-DFA and multi-peak analysis, the multifractality of LNG spot freight rates was verified, and factor signals were established, including spectral decomposition components, peak positions, and long-range correlation strength. Using methods such as LSTM, LNG futures return predictions were made, incorporating indicators like the Hurst exponent to achieve strategy switching and window smoothing. The backtest results showed an annualized return rate exceeding 36%.

Zero-shot OCT segmentation for volumetric and live-streaming data

Published Paper

Sep 2023 - Feb 2024, Published in SPIE

Authors: Kaixiang Zhang, Haoran Zhang, Qi Lan, Zesen Chen, Shuo Yin, Yongqi Wei, Zehao Wang, Jianlong Yang

This research presents a zero-shot OCT segmentation method that overcomes the limitations of traditional deep learning approaches, which require re-labeling and retraining for unseen Regions of Interest (ROIs). The method only needs box or point prompts for ROIs in the first frame, enabling automatic segmentation in subsequent frames.

OCT Method OCT Results

Kaggle: Optiver Trading at the Close

Competition Link

Sep 2023 - Mar 2024, Bronze Medalist

  • Derived high-frequency market microstructure factors from order book and auction data, including buy-sell imbalance ratios, price-volume triple barriers, reference price lags/differentials, statistical moments, depth-weighted spreads, and volatility proxies
  • Built LightGBM model with rolling training and Optuna-based hyperparameter optimization; achieved MAE of 5.473 on validation set

BERT-based ESG NLP Scoring System

Project Report

Jul 2022 - Feb 2023, SJTU Outstanding PRP Project

Built a real-time database by web scraping public opinions and disclosure information from trading platforms using proxy pools to enhance crawler efficiency. Fine-tuned BERT NLP model to establish a text sentiment prediction framework, providing ESG ratings for 3,000+ listed companies based on mainstream ESG rating indicators

Additional Information

Language

  • English: TOEFL 108
  • Chinese: Native Speaker

Interests

  • Go (Weiqi): I have been a player of Go since childhood. AlphaGo is one of my motivations to study AI.
  • Games: Fan of strategic games including Texas Hold’em poker, Avalon (board game), and Balatro.
  • Sports: Enjoy various sports, play basketball and work out daily, currently learning tennis. Always happy to connect with workout buddies or sports partners! :)