Haidar Khan

Research Scientist (Meta, SDAIA, Amazon). CTO and co-founder (ai.astrolabe). Partner (KhanTeT).

profile.jpg

Kingston, New York 12401

My research interests cover a broad range of topics in NLP and machine learning. My recent work is mainly focused around understanding how to build efficient (small/fast) models efficiently (without breaking the compute bank) for NLP tasks. I am also excited about ZeroSumEval, a project we started to scale evaluation of LLMs with compute and capabilties.

I’ve had the privilege of pursuing this work with amazing colleages at industry and academia. Most recently, I was granted Premium Residency for special talent in Saudi Arabia and spent a year with the National Center of AI (NCAI) at the Saudi Data and AI Authority (SDAIA) as a research scientist. Our team published research (1, 2, 3) on large language models (LLM) and built ALLaM, the best LLM for Arabic and English (at the time).

Prior to that, I was a senior applied scientist at Amazon Alexa AI conducting research in NLP topics (semantic parsing, efficient modeling) and helped build Alexa Teacher Models, Amazon’s precursor to Amazon General Intelligence.

I completed my PhD with the Data Science Research Center (DSRC) at Rensselaer Polytechnic Institute (RPI). My work with Prof. Bülent Yener focused on developing methods specially designed for classifying, predicting, and analyzing medical data. I was fortunate to work closely with a team of doctors from the Mount Sinai Hospital Epilepsy Center and witness the impact of my work on patient care.

The rest of my time is dedicated to my family and pursuing my other passions; learning (human) langauges (4 so far), Quran and Arabic poetry, riding horses (western), and hunting.

news

Sep 30, 2024 We founded ai.astrolabe
Aug 21, 2023 I joined the National Center for Artificial Intelligence in Riyadh, Saudi Arabia as a visiting scientist.

latest posts

Oct 08, 2024 A journey into NLP

selected publications

  1. Learning filter widths of spectral decompositions with wavelets
    Haidar Khan, and Bulent Yener
    Advances in Neural Information Processing Systems, 2018
  2. Don’t Parse, Insert: Multilingual Semantic Parsing with Insertion Based Decoding
    Qile Zhu, Haidar Khan, Saleh Soltan, and 2 more authors
    arXiv preprint arXiv:2010.03714, 2020
  3. Compressing Transformer-Based Semantic Parsing Models using Compositional Code Embeddings
    Prafull Prakash, Saurabh Kumar Shashidhar, Wenlong Zhao, and 3 more authors
    arXiv preprint arXiv:2010.05002, 2020
  4. RescoreBERT: Discriminative Speech Recognition Rescoring With Bert
    Liyan Xu, Yile Gu, Jari Kolehmainen, and 5 more authors
    In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022
  5. Alexa teacher model: Pretraining and distilling multi-billion-parameter encoders for natural language understanding systems
    Jack FitzGerald, Shankar Ananthakrishnan, Konstantine Arkoudas, and 8 more authors
    In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2022
  6. AlexaTM 20B: Few-Shot Learning Using a Large-Scale Multilingual Seq2Seq Model
    Saleh Soltan, Shankar Ananthakrishnan, Jack FitzGerald, and 8 more authors
    arXiv preprint arXiv:2208.01448, 2022
  7. When Benchmarks are Targets: Revealing the Sensitivity of Large Language Model Leaderboards
    Norah Alzahrani, Hisham Abdullah Alyahya, Yazeed Alnumay, and 8 more authors
    arXiv preprint arXiv:2402.01781, 2024
  8. A Systematic Survey and Critical Review on Evaluating Large Language Models: Challenges, Limitations, and Recommendations
    Md Tahmid Rahman Laskar, Sawsan Alqahtani, M Saiful Bari, and 8 more authors
    arXiv preprint arXiv:2407.04069, 2024
  9. ALLaM: Large Language Models for Arabic and English
    M Saiful Bari, Yazeed Alnumay, Norah A Alzahrani, and 8 more authors
    arXiv preprint arXiv:2407.15390, 2024