-
Notifications
You must be signed in to change notification settings - Fork 4.8k
ON-prem SQL databse #2493
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
For connecting to your on-premises SQL Server, I'd suggest using Azure Data Factory with a self-hosted Integration Runtime
|
|
The pipeline in Azure Data Factory is designed to handle data processing automatically. Here's how it works - first, we set up the pipeline structure using this code:
resource pipeline 'Microsoft.DataFactory/factories/pipelines@2018-06-01' = {
name: '${dataFactory.name}/DataPipeline'
properties: {
activities: [
{
name: 'CopyFromSource'
type: 'Copy'
inputs: [{ referenceName: 'SourceDataset', type: 'DatasetReference' }]
outputs: [{ referenceName: 'BlobDataset', type: 'DatasetReference' }]
}
]
}
}
@app.route(route="process_data")
def process_data(req: func.HttpRequest):
# Process documents with BERT vectorization
vectorizer = TextVectorizer()
for record in data:
vector = vectorizer.vectorize(process_text(record))
record['vector'] = vector
# Store processed results
blob_client.upload_data({'processed_data': processed_data}) The pipeline runs on a scheduled recurrence, automatically handling the entire process from data extraction through vectorization to storage in blob storage. When new data arrives, the function processes it using BERT vectorization and stores both original and processed versions. |
why cant i use the connection string to directly connect to my on prem server rather then have it on cloud and set up a pipeline to refresh ? |
You can definitely connect directly using a SQL connection string, but I recommend Azure Data Factory because this OpenAI demo needs extra processing like vectorization and indexing—which ADF securely automates without needing you to manage firewall rules, VPNs, or custom pipelines |
How can we connect this solution to an on prem SQL server so my data is directly pulled from there
The text was updated successfully, but these errors were encountered: