Hiroto Kurita

Hi there 👋 I am a 2nd-year M.S. student at TohokuNLP Group, Tohoku University, working with Kentaro Inui and Sho Yokoi.
🧪 Working on theoretical understanding of text embeddings and large language models, also enjoys training large models! 🤖
🤝 LinkedIn / Twitter / hiroto.kurita.q4 [at] dc.tohoku.ac.jp
📝Blogs2023/10/25 6:342023/10/25 12:17
 

Education

October 2022- Present, Sendai
 
Tohoku University M.S. in Information Science
Advisor: Kentaro Inui, Sho Yokoi
April 2019- Sept. 2022, Sendai
 
Tohoku University B.E. in Computer Science
Advisor: Kentaro Inui, Sho Yokoi
Early Graduation
 

Work Experiences

October 2023- Present, Tokyo
Research & Machine Learning Engineer Intern Kotoba Technologies, Inc.
October 2023- Present, Sendai
 
Research Assistant Tohoku University
October 2023- Present, Sendai
Research Assistant Tohoku University
April 2023- September 2023, Sendai
Teaching Assistant Tohoku University
Information and data literacy
July 2022- April 2023, Tokyo
Machine Learning Engineer Intern ELYZA, Inc.
Worked on NLP-related projects [interview article]
April 2019- September 2022, Sendai
Machine Learning Software Engineer Intern Adansons, Inc.
Working on data-centric AI platform products Mentored two student interns from India
August 2022, Tokyo
Summer Intern Nikkei
September 2020- Mar. 2021, Atlanta Online
 
Research Intern Georgia Institute of Technology Advisor: Biagio Mandracchia

Projects

Jan. 2023- Present
 
Training Japanese Large Language Models with Massive Scale CPUs on Fugaku (GPT-Fugaku)
Main collaborators: Keisuke Sakaguchi, Rio Yokota, Noriyuki Kojima, Jungo Kasai, Shota Sasaki, Kazuki Fujii, Shukai Nakamura, Taishi Nakamura, and more.
May 2023
Three-Points Scale Up of Japanese Large Language Models Training
2023 1st ABCI Grand Challenge Collaborators: Keisuke Sakaguchi, Rio Yokota, Koji Nishiguchi, Noriyuki Kojima, Jungo Kasai, Shota Sasaki, Kazuto Ando and Shukai Nakamura

Publications (Refereed)

 
  • Hiroto Kurita, Goro Kobayashi, Sho Yokoi and Kentaro Inui. ”Contrastive Learning-based Sentence Encoders Implicitly Weight Informative Words” In findings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP 2023), December 2023. [paper] [arXiv] [github]

Publications (Non-refereed)

  • 栗田 宙人, 小林 悟郎, 横井 祥, 乾 健太郎. BERTを用いた文埋め込みモデルによる単語の暗黙的な重み付け. 言語処理学会第29回年次大会, March 2023.

Awards

Talks

  • Hiroto Kurita. Recent Trends of Large Language Models. WBA Tohoku Summer Meetup (Invited Talk). July 2023.

Misc.

  • TOEFL iBT 98/120pts 2021
  • TOEIC L&R 960/990pts 2021