Age, Biography and Wiki

Quoc V. Le (Lê Viết Quốc) was born on 1982 in Hương Thủy, Thừa Thiên Huế, Vietnam, is a Vietnamese-American computer scientist. Discover Quoc V. Le's Biography, Age, Height, Physical Stats, Dating/Affairs, Family and career updates. Learn How rich is he in this year and how he spends money? Also learn how he earned most of networth at the age of 42 years old?

Popular As Lê Viết Quốc
Occupation N/A
Age 42 years old
Zodiac Sign N/A
Born
Birthday
Birthplace Hương Thủy, Thừa Thiên Huế, Vietnam
Nationality Vietnam

We recommend you to check the complete list of Famous People born on . He is a member of famous Computer with the age 42 years old group.

Quoc V. Le Height, Weight & Measurements

At 42 years old, Quoc V. Le height not available right now. We will update Quoc V. Le's Height, weight, Body Measurements, Eye Color, Hair Color, Shoe & Dress size soon as possible.

Physical Status
Height Not Available
Weight Not Available
Body Measurements Not Available
Eye Color Not Available
Hair Color Not Available

Dating & Relationship status

He is currently single. He is not dating anyone. We don't have much information about He's past relationship and any previous engaged. According to our Database, He has no children.

Family
Parents Not Available
Wife Not Available
Sibling Not Available
Children Not Available

Quoc V. Le Net Worth

His net worth has been growing significantly in 2023-2024. So, how much is Quoc V. Le worth at the age of 42 years old? Quoc V. Le’s income source is mostly from being a successful Computer. He is from Vietnam. We have estimated Quoc V. Le's net worth, money, salary, income, and assets.

Net Worth in 2024 $1 Million - $5 Million
Salary in 2024 Under Review
Net Worth in 2023 Pending
Salary in 2023 Under Review
House Not Available
Cars Not Available
Source of Income Computer

Quoc V. Le Social Network

Instagram
Linkedin
Twitter
Facebook
Wikipedia
Imdb

Timeline

1982

Lê Viết Quốc (born 1982), or in romanized form Quoc Viet Le, is a Vietnamese-American computer scientist and a machine learning pioneer at Google Brain, which he established with others from Google.

He co-invented the doc2vec and seq2seq models in natural language processing.

Le also initiated and lead the AutoML initiative at Google Brain, including the proposal of neural architecture search.

Le was born in Hương Thủy in the Thừa Thiên Huế province of Vietnam.

He studied at Quốc Học Huế High School.

2004

In 2004, Le moved to Australia and attended Australian National University for Bachelor's program, during which he worked under Alex Smola on Kernel method in machine learning.

2007

In 2007, Le moved to Stanford University for graduate studies in computer science, where his PhD advisor was Andrew Ng.

2011

In 2011, Le became a founding member of Google Brain along with his then PhD advisor Andrew Ng, Google Fellow Jeff Dean and Google researcher Greg Corrado.

Le led Google Brain's first major discovery, a deep learning algorithm trained on 16,000 CPU cores, which learned to recognize cats after watching only YouTube videos, and without ever having been told what a "cat" is.

2014

In 2014, Ilya Sutskever, Oriol Vinyals and Le proposed the seq2seq model for machine translation.

In the same year, Tomáš Mikolov and Le proposed the doc2vec model for representation learning of documents.

Le is among the lead authors and researchers of Google Neural Machine Translation.

Le initiated and lead the AutoML project at Google Brain, including the proposal of neural architecture search.

Le was named MIT Technology Review's innovators under 35 in 2014.

He has been interviewed by and his research has been reported in major media outlets including Wired, the New York Times, the Atlantic, and the MIT Technology Review.

Le was named an Alumni Laureate of the Australian National University School of Computing in 2022.

2020

Le is among the authors of LaMDA, a conversational large language model, originally developed and introduced as Meena in 2020.

In 2022, Le and co-authors proposed chain-of-thought prompting as a method to improve the reasoning ability of large language models.