Open in app

Sign In

Write

Sign In

Ajit Rajasekharan
Ajit Rajasekharan

777 Followers

Home

About

Pinned

ChatGPT — an epochal event

A quiet shift has been in progress over the last few years in the NLP ecosystem driven by large language models(LLMs). This culminated in…

AI

12 min read

ChatGPT — an epochal event
ChatGPT — an epochal event
AI

12 min read


Published in Towards Data Science

·Jun 29, 2022

A Quiet Shift in the NLP Ecosystem

It is already in motion… — The gist A paper published a few days ago (15 June 2022) offers a key insight into the impressive task-solving capabilities of the recent spate of billion+ parameter language models. Large language models exihibit the ability to solve specific tasks when two conditions are met — the model parameter size exceeds 100…

Artificial Intelligence

18 min read

A Quiet Shift in the NLP Ecosystem
A Quiet Shift in the NLP Ecosystem
Artificial Intelligence

18 min read


Published in Towards Data Science

·Mar 30, 2022

Almost Any Image Is Only 8k Vectors

An image representation that’s a close analog for words in generative tasks — Overview The search for an answer to the question “what is the analog of words in images?” appears to be broadly proceeding along two paths (with a few exceptions) driven by multiple factors — the nature of the task being solved (discriminative or generative ) how a model is trained (self-supervised…

Machine Learning

23 min read

Almost Any Image Is Only 8k Vectors
Almost Any Image Is Only 8k Vectors
Machine Learning

23 min read


Published in Towards Data Science

·Jan 24, 2022

SSL could Avoid Supervised Learning

For select supervised tasks with self-supervised learning(SSL) models satisfying certain properties — TL;DR Self-supervised learning (SSL) could be used to avoid supervised learning for some tasks leveraging self-supervised models like BERT, as is, without fine-tuning (supervision). For instance, this post describes an approach to perform named-entity recognition without fine-tuning a model on sentences. Instead, a small subset of BERT’s learned vocabulary is manually…

Naturallanguageprocessing

40 min read

SSL could Avoid Supervised Learning
SSL could Avoid Supervised Learning
Naturallanguageprocessing

40 min read


Published in Towards Data Science

·Apr 10, 2021

Quantitative evaluation of a pre-trained BERT model

A prerequisite to use a pre-trained model as is, without fine tuning — TL;DR Self-supervised learning is being leveraged off at scale using transformers, not only for text, but lately also for images(CLIP, ALIGN), to solve traditionally supervised tasks (e.g. classification), either as is, or with subsequent fine tuning. …

Naturallanguageprocessing

10 min read

Quantitative evaluation of a pre-trained BERT model
Quantitative evaluation of a pre-trained BERT model
Naturallanguageprocessing

10 min read


Published in Towards Data Science

·Jan 9, 2021

Swiss army knife for unsupervised task solving

BERT is a prize addition to the practitioner’s toolbox — TL;DR Natural language processing tasks traditionally requiring labeled data could be solved entirely or in part, subject to a few constraints, without the need for labeled data by leveraging the self-supervised learning of a BERT model, provided those tasks lend themselves to be viewed entirely or in part, as a similarity…

Naturallanguageprocessing

14 min read

Swiss army knife for unsupervised task solving
Swiss army knife for unsupervised task solving
Naturallanguageprocessing

14 min read


Published in Towards Data Science

·Dec 18, 2020

Unsupervised synonym harvesting

A hybrid approach combining symbolic processing with distributed representations — TL;DR Extracting all the different ways a particular term can be referred to (synonym harvesting) is key for applications in biomedical domain where drugs, genes etc. have many synonyms. While there are human curated knowledge bases for synonyms in the biomedical domain, they are generally incomplete, continually trying to play catchup…

Naturallanguageprocessing

11 min read

Unsupervised synonym harvesting
Unsupervised synonym harvesting
Naturallanguageprocessing

11 min read


Published in Towards Data Science

·Nov 4, 2020

Maximizing BERT model performance

An approach to evaluate a pre-trained BERT model to increase performance — TL;DR Training a BERT model from scratch on a domain specific corpus such as biomedical space with a custom vocabulary generated specific to that space has proven to be critical to maximize model performance in biomedical domain. This is largely because of language characteristics that are unique to biomedical space which…

NLP

14 min read

Maximizing BERT model performance
Maximizing BERT model performance
NLP

14 min read


Oct 12, 2020

A hope infused narrative to course correct our planet from imminent ruin

The data driven argument David Attenborough makes about our planet’s future, using his own lifetime as the scale of measure, in the recently released film on Netflix “A life on our planet”, is hard to ignore, despite the fatigue that may have desensitized us to this topic by the relentless…

Climate Change

3 min read

A hope infused narrative to course correct our planet from imminent ruin
A hope infused narrative to course correct our planet from imminent ruin
Climate Change

3 min read


Published in Towards Data Science

·Jul 5, 2020

Unsupervised creation of interpretable sentence representations

For sentence similarity/document search applications — TL;DR To date, models learn fixed size representation of sentences, typically with some form of supervision, which are then used for sentence similarity or other downstream tasks. Examples of this are Google’s Universal sentence encoder (2018) and Sentence transformers (2019). Supervised learning of fixed size representations tends to outperform unsupervised creation…

Naturallanguageprocessing

10 min read

Unsupervised creation of interpretable sentence representations
Unsupervised creation of interpretable sentence representations
Naturallanguageprocessing

10 min read

Ajit Rajasekharan

Ajit Rajasekharan

777 Followers

Machine learning practitioner

Following
  • Peter Barrett Bryan

    Peter Barrett Bryan

  • Sik-Ho Tsang

    Sik-Ho Tsang

  • elvis

    elvis

  • Sergi Castella i Sapé

    Sergi Castella i Sapé

  • Ben Huberman

    Ben Huberman

Help

Status

Writers

Blog

Careers

Privacy

Terms

About

Text to speech