SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
This is a sentence-transformers model finetuned from sentence-transformers/all-MiniLM-L6-v2. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
Model Details
Model Description
- Model Type: Sentence Transformer
- Base model: sentence-transformers/all-MiniLM-L6-v2
- Maximum Sequence Length: 128 tokens
- Output Dimensionality: 384 tokens
- Similarity Function: Cosine Similarity
Model Sources
- Documentation: Sentence Transformers Documentation
- Repository: Sentence Transformers on GitHub
- Hugging Face: Sentence Transformers on Hugging Face
Full Model Architecture
SentenceTransformer(
(0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
(1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
)
Usage
Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("Daxtra/sbert-summaries-minilm-24-batch")
# Run inference
sentences = [
"- Data Management Specialist (Junior/Intermediate), requiring a Bachelor's degree in computer science, information systems, or related field.\n- Must have a solid understanding of web- and app-based platforms.\n- Required experience with SQL, Data warehousing, and Tableau.\n- Ability to analyze, interpret, and organize large data sets.\n- Proficiency in Microsoft Suite, G-Suite, Slack, Zoom, ZenDesk, and Monday.com is essential.",
"- Leadership roles in healthcare and pharmaceuticals with extensive experience in contract negotiation and site management.\n- Parexel Site Contract Leader: Lead global CSA strategy development, budget creation, and legal drafting, ensuring compliance with ICH-GCP and local regulations.\n- Drafts, reviews, and finalizes contracts with study sites, managing budgets and negotiations, and maintaining quality standards.\n- Expertise in financial risk assessment, budgeting, and regulatory compliance with a Bachelor's in Criminal Justice.\n- Proficient in Microsoft Excel, PowerPoint, and Salesforce.com, with a background in biotechnology, clinical trials, and patient safety.",
"- Experienced Technical Project Manager with 12 years in IT infrastructure and project management.\n- Expertise in IT project planning, budgeting, resource allocation, and stakeholder engagement.\n- Led successful IT infrastructure upgrades, including EFT and Manta application migrations to AWS.\n- Managed cloud and on-premise infrastructure projects, ensuring compliance and minimal disruption.\n- Holds PMP certification and proficiency in Agile and Waterfall methodologies.\n- Master's in Business Administration with a focus on Marketing and User Experience.\n- Strong analytical skills, stakeholder management, and continuous improvement practices.\n- Proficient in Microsoft Office, Windows Server, and Azure, with experience in networking, security, and VMware.",
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 384]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
Evaluation
Metrics
Information Retrieval
- Dataset:
vac-res-matcher
- Evaluated with
InformationRetrievalEvaluator
Metric | Value |
---|---|
cosine_accuracy@10 | 0.3697 |
cosine_precision@10 | 0.0737 |
cosine_recall@10 | 0.1092 |
cosine_ndcg@10 | 0.1177 |
cosine_mrr@10 | 0.1953 |
cosine_map@10 | 0.0713 |
Training Details
Training Dataset
Unnamed Dataset
- Size: 149,352 training samples
- Columns:
sentence_0
andsentence_1
- Approximate statistics based on the first 1000 samples:
sentence_0 sentence_1 type string string details - min: 53 tokens
- mean: 115.57 tokens
- max: 128 tokens
- min: 51 tokens
- mean: 119.82 tokens
- max: 128 tokens
- Samples:
sentence_0 sentence_1 - Solutions Architect/Snowflake position requiring 8+ years of experience in data management and implementation of Snowflake solutions.
- Responsibilities include developing data models, optimizing data pipelines, and integrating with AWS/Azure PBAS.
- Must possess deep knowledge of relational and NoSQL databases, SQL, and Unix Shell/Python Scripting.
- Experience with data security, database optimization, and Snowflake's Resource Monitors.
- Required: Bachelor's or Master's degree in Computer Science or related field.
- Preferred: Management consulting experience, expertise in AI use cases, and experience with Cloud technologies.
- Skills: Strong leadership, team management, and experience with globally distributed teams.- Senior IT Consultant with 30 years of experience, specializing in enterprise architecture and solution architecture.
- Currently a Senior Consultant, with significant experience in architecture roles and delivering solutions for mission-critical projects.
- Expertise in Enterprise Architecture, Enterprise Consortium management, and IT Governance.
- Leads projects ranging from $100M to $900M, focusing on architecture for business, data, and applications.
- Designed enterprise-wide architecture models and created strategies for product development based on product architecture.
- Skilled in modern technologies like Spring Framework, Kafka, Spring Cloud, and Docker, and in containerization with Kubernetes.
- Holds Masters Degrees in Computer Science and Law, with a Bachelor's in Computer Sciences.- Global Head of Business Development - Financial Education, with a focus on rapid adaptation and leadership in Hong Kong.
- Responsible for identifying leads, building pipelines, and converting prospects into clients, and managing international teams.
- Implement relationship-based sales practices, nurture industry relationships, and lead global expansion efforts.
- Requires a proven track record in financial services business development and experience in sales with tech, media, or administration.
- Strong networking and relationship-building skills, leadership abilities across cultures, and excellent communication and negotiation skills.
- Must have the right to work in the United Kingdom.- Management Accountant with extensive experience in financial planning and analysis, specializing in the banking sector.
- Leads financial planning and analysis roles, overseeing budget processes, regulatory reporting, and financial performance analysis.
- Streamlined processes to reduce payment processing time by 15% and minimize penalties.
- Led quarterly and year-end audits, ensuring accurate financial audits and implementing improvement suggestions.
- Proficient in financial reporting, budgeting, forecasting, and variance analysis; adept at using Microsoft Excel and QuickBooks.
- Strong communication and team management skills, with expertise in process improvement and strategic initiatives.- Customer Service Administrator position requiring experience in customer service or administration.
- Key responsibilities include managing customer feedback, coordinating deliveries, processing refunds, and general office support.
- Must possess strong communication, problem-solving, and attention to detail skills.
- Proficiency in MS Office and CRM software required.
- Ability to multitask and prioritize workload effectively.
- Must have the right to work in the United Kingdom.- Experienced Technical Services Coordinator with a strong background in administration and team collaboration, seeking to advance in London.
- Manages UK technical services, handling device delivery, fault reporting, and resource planning.
- Expert in Microsoft Office, maintaining databases and customer service through communication and negotiation.
- Holds NVQ Level 2 and 3, with skills in GDPR compliance and financial transactions.
- Proven experience in collecting evidence for claims, negotiating settlements, and managing key performance indicators.
- Strong communication, organizational, and interpersonal skills, with proficiency in using initiative and effective negotiation.
- Holds GCSE qualifications and relevant experience in customer service roles. - Loss:
MultipleNegativesRankingLoss
with these parameters:{ "scale": 20.0, "similarity_fct": "cos_sim" }
Training Hyperparameters
Non-Default Hyperparameters
eval_strategy
: stepsper_device_train_batch_size
: 24per_device_eval_batch_size
: 24num_train_epochs
: 1multi_dataset_batch_sampler
: round_robin
All Hyperparameters
Click to expand
overwrite_output_dir
: Falsedo_predict
: Falseeval_strategy
: stepsprediction_loss_only
: Trueper_device_train_batch_size
: 24per_device_eval_batch_size
: 24per_gpu_train_batch_size
: Noneper_gpu_eval_batch_size
: Nonegradient_accumulation_steps
: 1eval_accumulation_steps
: Nonetorch_empty_cache_steps
: Nonelearning_rate
: 5e-05weight_decay
: 0.0adam_beta1
: 0.9adam_beta2
: 0.999adam_epsilon
: 1e-08max_grad_norm
: 1num_train_epochs
: 1max_steps
: -1lr_scheduler_type
: linearlr_scheduler_kwargs
: {}warmup_ratio
: 0.0warmup_steps
: 0log_level
: passivelog_level_replica
: warninglog_on_each_node
: Truelogging_nan_inf_filter
: Truesave_safetensors
: Truesave_on_each_node
: Falsesave_only_model
: Falserestore_callback_states_from_checkpoint
: Falseno_cuda
: Falseuse_cpu
: Falseuse_mps_device
: Falseseed
: 42data_seed
: Nonejit_mode_eval
: Falseuse_ipex
: Falsebf16
: Falsefp16
: Falsefp16_opt_level
: O1half_precision_backend
: autobf16_full_eval
: Falsefp16_full_eval
: Falsetf32
: Nonelocal_rank
: 0ddp_backend
: Nonetpu_num_cores
: Nonetpu_metrics_debug
: Falsedebug
: []dataloader_drop_last
: Falsedataloader_num_workers
: 0dataloader_prefetch_factor
: Nonepast_index
: -1disable_tqdm
: Falseremove_unused_columns
: Truelabel_names
: Noneload_best_model_at_end
: Falseignore_data_skip
: Falsefsdp
: []fsdp_min_num_params
: 0fsdp_config
: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}fsdp_transformer_layer_cls_to_wrap
: Noneaccelerator_config
: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}deepspeed
: Nonelabel_smoothing_factor
: 0.0optim
: adamw_torchoptim_args
: Noneadafactor
: Falsegroup_by_length
: Falselength_column_name
: lengthddp_find_unused_parameters
: Noneddp_bucket_cap_mb
: Noneddp_broadcast_buffers
: Falsedataloader_pin_memory
: Truedataloader_persistent_workers
: Falseskip_memory_metrics
: Trueuse_legacy_prediction_loop
: Falsepush_to_hub
: Falseresume_from_checkpoint
: Nonehub_model_id
: Nonehub_strategy
: every_savehub_private_repo
: Falsehub_always_push
: Falsegradient_checkpointing
: Falsegradient_checkpointing_kwargs
: Noneinclude_inputs_for_metrics
: Falseeval_do_concat_batches
: Truefp16_backend
: autopush_to_hub_model_id
: Nonepush_to_hub_organization
: Nonemp_parameters
:auto_find_batch_size
: Falsefull_determinism
: Falsetorchdynamo
: Noneray_scope
: lastddp_timeout
: 1800torch_compile
: Falsetorch_compile_backend
: Nonetorch_compile_mode
: Nonedispatch_batches
: Nonesplit_batches
: Noneinclude_tokens_per_second
: Falseinclude_num_input_tokens_seen
: Falseneftune_noise_alpha
: Noneoptim_target_modules
: Nonebatch_eval_metrics
: Falseeval_on_start
: Falseeval_use_gather_object
: Falsebatch_sampler
: batch_samplermulti_dataset_batch_sampler
: round_robin
Training Logs
Epoch | Step | Training Loss | vac-res-matcher_cosine_map@10 |
---|---|---|---|
0.0803 | 500 | 1.5369 | - |
0.1000 | 622 | - | 0.0697 |
0.1607 | 1000 | 1.2768 | - |
0.1999 | 1244 | - | 0.0692 |
0.2410 | 1500 | 1.2 | - |
0.2999 | 1866 | - | 0.0673 |
0.3214 | 2000 | 1.1463 | - |
0.3998 | 2488 | - | 0.0705 |
0.4017 | 2500 | 1.1206 | - |
0.4821 | 3000 | 1.1043 | - |
0.4998 | 3110 | - | 0.0683 |
0.5624 | 3500 | 1.0768 | - |
0.5997 | 3732 | - | 0.0700 |
0.6428 | 4000 | 1.0905 | - |
0.6997 | 4354 | - | 0.0705 |
0.7231 | 4500 | 1.0804 | - |
0.7996 | 4976 | - | 0.0699 |
0.8035 | 5000 | 1.0536 | - |
0.8838 | 5500 | 1.0352 | - |
0.8996 | 5598 | - | 0.0715 |
0.9642 | 6000 | 1.0292 | - |
0.9995 | 6220 | - | 0.0713 |
1.0 | 6223 | - | 0.0713 |
Framework Versions
- Python: 3.10.12
- Sentence Transformers: 3.2.1
- Transformers: 4.44.2
- PyTorch: 2.4.1+cu121
- Accelerate: 0.34.2
- Datasets: 3.0.1
- Tokenizers: 0.19.1
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
MultipleNegativesRankingLoss
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
- Downloads last month
- 4
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for Daxtra/sbert-summaries-minilm-24-batch
Base model
sentence-transformers/all-MiniLM-L6-v2Evaluation results
- Cosine Accuracy@10 on vac res matcherself-reported0.370
- Cosine Precision@10 on vac res matcherself-reported0.074
- Cosine Recall@10 on vac res matcherself-reported0.109
- Cosine Ndcg@10 on vac res matcherself-reported0.118
- Cosine Mrr@10 on vac res matcherself-reported0.195
- Cosine Map@10 on vac res matcherself-reported0.071