<div style="line-height:1.7;color:#000000;font-size:14px;font-family:Arial"><div>Hi Adapters,</div><div><p> Recent
years have seen the remarkable progress of large language models
by scaling up data size and model size,
these LLMs raise amazing emergent abilities, typically including In-Context
Learning (ICL), instruction following, and Chain of Thought (CoT).
Although
LLMs have demonstrated surprising zero/few-shot reasoning performance on most
Natural Language Processing (NLP) tasks, they are inherently "blind"
to vision since they can only understand discrete text. </p><p> Today, i'll follow last week xukai's seminar, i'll give a brief introduction on vision-language tasks and datasets, as well as the structure of the mainstream vision large language models.</p><pre> Hope you find this talk interesting!
Time: Wed 10 am. - 11:30 am.
Meeting link:
https://teams.microsoft.com/l/meetup-join/19%3ameeting_M2VmMTU5MzgtODUzOC00NmU4LTg0MzktNGFjNDdiMmIwYTI1%40thread.v2/0?context=%7b%22Tid%22%3a%225cdc5b43-d7be-4caa-8173-729e3b0a62d9%22%2c%22Oid%22%3a%221a8b9fa0-af57-4a1c-9390-22d1c201d622%22%7d
Best wishes,
Apple</pre></div></div>