Not at all. I should have made clear how I extracted in the post.

I extracted embeddings from a pytorch model (pytorch_model.bin file). The code to extract is pasted here. It assumes the embeddings are stored with the name bert.embeddings.word_embeddings.weight. You can just print out all the keys in your clinicalbert model and see what the key name exactly is.

import torch

import pdb

md = torch.load(“./pytorch_model.bin”,map_location=’cpu’)

for k in md:

if (k == “bert.embeddings.word_embeddings.weight”):

embeds = md[k]

for l in range(len(embeds)):

vector = embeds[l]

for m in range(len(vector)):

print(round(vector[m].tolist(),6),end=’ ‘)





Machine learning practitioner

Love podcasts or audiobooks? Learn on the go with our new app.

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Ajit Rajasekharan

Ajit Rajasekharan

Machine learning practitioner

More from Medium

Dear God,

Hello, it’s me!

Data-Driven Decision Making in Public Policy