Not at all. I should have made clear how I extracted in the post.

I extracted embeddings from a pytorch model (pytorch_model.bin file). The code to extract is pasted here. It assumes the embeddings are stored with the name bert.embeddings.word_embeddings.weight. You can just print out all the keys in your clinicalbert model and see what the key name exactly is.

import torch

import pdb

md = torch.load(“./pytorch_model.bin”,map_location=’cpu’)

for k in md:

if (k == “bert.embeddings.word_embeddings.weight”):

embeds = md[k]

for l in range(len(embeds)):

vector = embeds[l]

for m in range(len(vector)):

print(round(vector[m].tolist(),6),end=’ ‘)


Machine learning practitioner

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store