Jump to content
Main menu
Main menu
move to sidebar
hide
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Special pages
Large Language Model Wiki
Search
Search
Appearance
Create account
Log in
Personal tools
Create account
Log in
Pages for logged out editors
learn more
Contributions
Talk
Editing
Ashish Vaswani
Page
Discussion
English
Read
Edit
View history
Tools
Tools
move to sidebar
hide
Actions
Read
Edit
View history
General
What links here
Related changes
Page information
Appearance
move to sidebar
hide
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{Infobox scientist | name = Ashish Vaswani | image = [[File:Ashish Vaswani portrait.jpg|250px]] <!-- Replace with actual image if available --> | caption = Ashish Vaswani | birth_date = 1986 | birth_place = India | citizenship = Indian | fields = [[Computer science]] β’ [[Artificial intelligence]] β’ [[Machine learning]] β’ [[Natural language processing]] | workplaces = Google Brain<br />Adept AI<br />Essential AI | alma_mater = Birla Institute of Technology, Mesra (B.E.)<br />University of Southern California (M.S., Ph.D.) | doctoral_advisor = Liang Huang<br />David Chiang | known_for = Co-author of "[[Attention Is All You Need]]"<br />[[Transformer (deep learning architecture)|Transformer]] architecture | awards = Best Paper Award, Information Sciences Institute Graduate Research Symposium (2010)<br />S. Chandrasekhar Rising Indian Diaspora Scientist Award }} '''Ashish Vaswani''' (born 1986) is an Indian computer scientist and artificial intelligence researcher. He is best known as the first author of the seminal 2017 paper "[[Attention Is All You Need]]", which introduced the [[Transformer (deep learning architecture)|Transformer]] neural network architecture. This architecture has become the foundational building block for nearly all modern large language models (LLMs), including those powering systems like ChatGPT, BERT, and many others. Vaswani is the co-founder and CEO of Essential AI, a company focused on building open, powerful AI systems to solve complex real-world challenges. == Early life and education == Vaswani was born in India in 1986. He spent part of his childhood in Oman before his family moved to Nagpur when he was 15. He developed an early interest in science and mathematics. He earned a Bachelor's degree in Computer Science and Engineering from the Birla Institute of Technology, Mesra. He later moved to the United States, completing a Master's degree and a Ph.D. in Computer Science at the University of Southern California (USC). His doctoral advisors were Liang Huang and David Chiang. During his Ph.D., he conducted research at the Information Sciences Institute at USC, with a focus on natural language processing and machine translation. == Career == After completing his Ph.D., Vaswani joined Google Brain, where he worked as a research scientist for more than six years. At Google, he contributed to advancements in natural language processing and deep learning. In 2017, while at Google Brain, Vaswani led the team that published "Attention Is All You Need." The paper proposed replacing recurrent and convolutional layers with a purely attention-based mechanism, enabling greater parallelism, faster training, and superior performance on sequence transduction tasks such as machine translation. In 2021β2022, Vaswani co-founded Adept AI with Niki Parmar and other colleagues, focusing on training neural networks to perform practical tasks and actions. He served as Co-Founder and Chief Scientist. In late 2022/early 2023, Vaswani and Parmar left Adept to found Essential AI. As CEO, Vaswani leads the company in developing frontier AI models with an emphasis on openness, collaboration, and solving humanity's biggest challenges through advanced reasoning systems. Essential AI has raised significant funding and collaborates on hardware platforms such as AMD Instinct GPUs. == Contributions == Vaswani's primary contribution is the Transformer architecture, which uses self-attention mechanisms to process entire sequences in parallel rather than sequentially. This breakthrough addressed key limitations of previous models (RNNs and LSTMs), dramatically improving scalability and performance. The Transformer has since been adapted for: * Natural language processing (NLP) * Computer vision (Vision Transformers) * Multimodal tasks * Image and music generation * Scientific applications (e.g., DNA sequence analysis) His earlier work during his Ph.D. included research on unsupervised word alignment for machine translation. As of 2026, Vaswani continues to advocate for open science approaches in AI development. == Awards and recognition == * Best Paper Award at the Information Sciences Institute Graduate Research Symposium (2010) * Best Paper Award at the 25th Army Science Conference (2006) * S. Chandrasekhar Rising Indian Diaspora Scientist Award The "Attention Is All You Need" paper has received over 200,000 citations, making it one of the most influential works in modern artificial intelligence. == Personal life == Vaswani is based in the San Francisco Bay Area. He maintains a relatively low public profile compared to many AI leaders, focusing on technical research and building teams for ambitious challenges. == See also == * [[Transformer (deep learning architecture)]] * [[Attention Is All You Need]] * [[Google Brain]] * [[Essential AI]] == References == {{Reflist}} [[Category:1986 births]] [[Category:Living people]] [[Category:Indian computer scientists]] [[Category:Artificial intelligence researchers]] [[Category:Machine learning researchers]] [[Category:Google Brain]] [[Category:Transformers (machine learning model)]] [[Category:University of Southern California alumni]]
Summary:
Please note that all contributions to Large Language Model Wiki may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see
My wiki:Copyrights
for details).
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Pages included on this page:
Template:Infobox scientist
(
edit
)
Template:Reflist
(
edit
)
Search
Search
Editing
Ashish Vaswani
Add topic