Skip to content
View yuyu5333's full-sized avatar
๐Ÿš‹
Studying
๐Ÿš‹
Studying

Block or report yuyu5333

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this userโ€™s behavior. Learn more about reporting abuse.

Report abuse
yuyu5333/README.md

Hi there ๐Ÿ‘‹ I'm Jay

๐Ÿ”ญ I'm currently exploring large language models and Diffusion Transformers (Dit), focusing on their training methodologies and inference optimization techniques.

๐ŸŽ“ I hold a Master's degree in Computer Science and Technology from Northwestern Polytechnical University, where I specialized in intelligent IoT terminals.

๐Ÿ‘ฏ I'm looking to collaborate on open source projects focused on machine learning, large model systems, and AI infrastructure.

๐Ÿ’ฌ Ask me about deep learning, large language models, Diffusion Transformers (Dit), Python programming, or FPGA-based system design.

๐Ÿ“ซ How to reach me: [email protected] or [email protected]

๐Ÿ˜„ Pronouns: He/Him

โšก Fun fact: I love tackling complex algorithms and have won several awards in national mathematics modeling competitions.

๐Ÿ… Some of my achievements include:

๐Ÿ‘จโ€๐Ÿ’ป My tech stack includes:

  • Programming Languages: C++, Python, Verilog
  • ML Frameworks: PyTorch, TensorFlow, MindSpore
  • Web Technologies: Django, Vue

๐ŸŒŸ I'm passionate about applying my skills to develop solutions that push the boundaries of technology and help improve our daily lives.

Pinned Loading

  1. VeOmni VeOmni Public

    Forked from ByteDance-Seed/VeOmni

    VeOmni: Scaling any Modality Model Training to any Accelerators with PyTorch native Training Framework

    Python

  2. flash-attention flash-attention Public

    Forked from Dao-AILab/flash-attention

    Fast and memory-efficient exact attention

    Python

  3. minimind minimind Public

    Forked from jingyaogong/minimind

    ๐Ÿš€๐Ÿš€ ใ€Œๅคงๆจกๅž‹ใ€2ๅฐๆ—ถๅฎŒๅ…จไปŽ0่ฎญ็ปƒ26M็š„ๅฐๅ‚ๆ•ฐGPT๏ผ๐ŸŒ Train a 26M-parameter GPT from scratch in just 2h!

    Python