Skip to content
View Pittawat2542's full-sized avatar

Block or report Pittawat2542

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please donโ€™t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this userโ€™s behavior. Learn more about reporting abuse.

Report abuse
Pittawat2542/README.md

Hi, I'm Pittawat Taveekitworachai ๐Ÿ‘‹

GitHub followers Website Email Google Scholar

AI researcher-builder working on language models, reasoning, post-training, evaluation, and agentic AI systems.

I study how language models behave, reason, fail, and improve โ€” then turn those insights into open models, evaluation tools, datasets, applications, talks, and writing.

Currently, my work focuses on:

  • ๐Ÿง  LLM reasoning & inference-time strategies
  • ๐Ÿงช Post-training, SFT/RFT, RLHF, and behavior shaping
  • ๐Ÿงญ Evaluation, robustness, hallucination, and structured output following
  • ๐Ÿ› ๏ธ Agentic AI, context engineering, RAG, and tool-using workflows
  • ๐ŸŒ Thai / Southeast Asian AI, open models, and low-resource language systems
  • ๐ŸŽจ Research ร— Engineering ร— Design โ€” making ideas understandable, usable, and useful

Selected Work

๐ŸŒช๏ธ Typhoon & Open Thai AI

I work on Typhoon, a family of Thai open models and applied AI systems, including reasoning models, medical reasoning, evaluation, OCR, ASR, and practical AI applications.

  • Typhoon T1 โ€” open Thai reasoning model research preview
  • Typhoon-Si Med-Thinking 4B โ€” medical reasoning model for ranked diagnoses

Explore more: Projects ยท Publications

๐Ÿ“š Research

My research asks questions like:

  • Can we shape model behavior through post-training rather than only prompting?
  • How do prior prompts affect reinforcement fine-tuning?
  • When does hallucination hurt โ€” and when can it unexpectedly help?
  • How should we evaluate reasoning models beyond surface-level benchmark scores?
  • How can we build more reliable AI systems for domains like medicine, finance, games, and Thai language tasks?

See: Google Scholar ยท ACL Anthology


Open Source & Tools

I enjoy building tools that help people evaluate, understand, and apply AI systems.

Some public projects include:

  • ๐Ÿงช BenchING โ€” structured output benchmark for LLMs
  • โš–๏ธ Themis โ€” lightweight platform for LLM evaluation experiments
  • ๐Ÿท๏ธ Thoth โ€” GUI tool for dataset labeling
  • ๐Ÿš— DriveSSD / ADAS distillation โ€” cost-effective LLM distillation for driving assessment
  • ๐Ÿ‡น๐Ÿ‡ญ Krathu-500 โ€” Thai post-comment corpus
  • ๐ŸŽฎ LLM story generation evaluation โ€” studying LLMs in narrative and game contexts

Browse more: github.com/Pittawat2542


Writing & Talks

I write and speak about AI research, LLM reasoning, post-training, agentic workflows, and the future of AI systems.

Recent topics include:

  • Agentic AI with Context Engineering
  • Open Models, Smarter Agents
  • How to Train Your Reasoning Models
  • From Text to Thought: How Modern AI Models Learn to Reason
  • Open-Source Typhoon: Democratizing Advanced AI
  • AI as the Common Language of Knowledge

Read more: Blog ยท Talks


What I Care About

Iโ€™m interested in AI systems that are not only powerful, but also:

  • reliable enough to evaluate honestly,
  • useful enough to solve real problems,
  • open enough to expand access,
  • and understandable enough for people to reason with.

My long-term goal is to help push the boundary of knowledge while building AI that meaningfully serves humanity.


๐Ÿ† GitHub Trophies

GitHub Trophies


Connect

Pinned Loading

  1. thoth thoth Public

    A GUI program helping for labeling dataset

    Python 4 2

  2. web-dev-crash-course web-dev-crash-course Public

    Web Development Crash Course for CS@SIT, KMUTT

    JavaScript 4 1

  3. themis themis Public

    Themis is a Python package for running reproducible LLM evaluations

    Python 3

  4. null-shot-prompting null-shot-prompting Public

    This repository contains the code for the paper "Null-Shot Prompting: Rethinking Prompting Large Language Models With Hallucination".

    Python 2

  5. thailand-earthquake-visualizer thailand-earthquake-visualizer Public

    An interactive visualization tool for earthquake data from Thailand and surrounding regions.

    JavaScript 2

  6. themis-examples themis-examples Public

    Examples for `themis`, an evaluation package