data-factory | 🏭Auto generate mock data for java test(便于 Java 测试自动生成对象信息) | Mock library
kandi X-RAY | data-factory Summary
kandi X-RAY | data-factory Summary
Auto generate mock data for java test.(便于 Java 测试自动生成对象信息)
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
- Build data value
- Get data class
- Get generic interfaces
- Init data factory
- Builds the data object
- Gets the data annotation
- Get data value
- Build context
- Build a String
- Get length of random length
- Validate parameters check
- Build a collection
- Create an Iterable instance
- Generate random int array
- Generate a random size
- Build a map
- Generate random float array
- Build short array
- Generate long array
- Build boolean array
- Generate a random double array
- Build char array
- Generate random byte array
- Build a random array
- Build a random enum
data-factory Key Features
data-factory Examples and Code Snippets
Community Discussions
Trending Discussions on data-factory
QUESTION
I'm trying to move some data from Azure SQL Server Database to Azure Blob Storage with the "Copy Data" pipeline in Azure Data Factory. In particular, I'm using the "Use query" option with the ?AdfDynamicRangePartitionCondition
hook, as suggested by Microsoft's pattern here, in the Source
tab of the pipeline, and the copy operation is parallelized by the presence of a partition key used in the query itself.
The source on SQL Server Database consists of two views with ~300k and ~3M rows, respectively. Additionally, the views have the same query structure, e.g. (pseudo-code)
...ANSWER
Answered 2021-Jun-10 at 06:24When there's a copy activity performance issue in ADF and the root cause is not obvious (e.g. if source is fast, but sink is throttled, and we know why) -- here's how I would go about it :
- Start with the Integration Runtime (IR) (doc.). This might be a jobs' concurrency issue, a network throughput issue, or just an undersized VM (in case of self-hosted). Like, >80% of all issues in my prod ETL are caused by IR-s, in one way or another.
- Replicate copy activity behavior both on source & sink. Query the views from your local machine (ideally, from a VM in the same environment as your IR), write the flat files to blob, etc. I'm assuming you've done that already, but having another observation rarely hurts.
- Test various configurations of copy activity. Changing
isolationLevel
,partitionOption
,parallelCopies
andenableStaging
would be my first steps here. This won't fix the root cause of your issue, obviously, but can point a direction for you to dig in further. - Try searching the documentation (this doc., provided by @Leon is a good start). This should have been a step #1, however, I find ADF documentation somewhat lacking.
N.B. this is based on my personal experience with Data Factory.
Providing a specific solution in this case is, indeed, quite hard.
QUESTION
Currently, I'm following this doc to use Oauth to copy data from Rest connector. I applied the suggested temple ,when I configure this web activity, as for the body content, it show I should provide below parameters. I wonder where to get this parameters?
...ANSWER
Answered 2021-Jun-11 at 05:46These are app registration ID and password. You need to register an app in Azure AD.
Below MSFT docs provides details about the same:
QUESTION
Currently I'm following this docs to use Oauth to copy data from REST connector into Azure Data Lake Storage in JSON format using OAuth. but I cannot find below temple in my temple gallery. I wonder where and how to get this temple ?
...ANSWER
Answered 2021-Jun-10 at 02:19As the following screenshot shows, you can click Add new resource button, then click pipeline from template and you will find it.
QUESTION
I have a path in ADLS that has a range of different files including *.csv
and *.xml
(which is not true xml
, it's just a csv
with xml
extension).
I want to copy only *.csv
and *.xml
files from this path to another using copy activity in ADF. Right now I only can specify one of them as wildcard in the file name of copy activity and not both.
Is there any way to specify two wildcards, like for example, .csv or .xml
.
BTW, I might be able to use filter activity with get meta data, but this is too much if there is other ways. This documentation didn't help much too:
As I said, filtering won't work (without forEach), and that's not optimized:
...ANSWER
Answered 2021-Jun-09 at 09:06No, there isn't a way can specify two wildcards path.
According my experience, the easiest way is that you can create two copy active in one pipeline:
- Copy active1: copy the files end with
*.csv
. - Copy active2: copy the files end with
*.xml
.
For your another question,there are many ways can achieve it. You could add an if condition to filter the condition: only copy active 1 and 2 both true/succeeded:
QUESTION
When using the "publish" on the Azure Data Factory the ARM Template is generated
...ANSWER
Answered 2021-Jun-08 at 20:33I am not able to reproduce the issue but would suggest not including the factory in the ARM template as documented here: https://docs.microsoft.com/en-us/azure/data-factory/author-global-parameters#cicd
Including the factory will cause other downstream issues when using the automated publish flow for CI/CD such as removing the git configuration on the source factory, so deploying global parameters with PowerShell is the recommended approach. By not including the factory in the ARM template, this error will not occur. Feel free to continue the discussion here: https://github.com/Azure/Azure-DataFactory/issues/285
QUESTION
I'm trying to set up a system where an Azure DataFactory can call an Azure function through its managed identity. Good example here: Authorising Azure Function App Http endpoint from Data Factory
However, this was using the old(er) Authentication/Authorization tool for Azure functions, which has now been renamed Authentication (Classic). Setting the system up through this is fine, I can make the call and get a response, but upgrading to Authorization causes this to break. It seems like the key thing missing is the option of "Action to take when the request is not authenticated", which I cannot seem to set with the new Authorization tool but should be set to "Login with Azure AD"
In summary, how do I set this setting with the new Authorization tool so that a MSI can make a cool to the function and authenticate with AAD.
Image with classic
Image with new Authorization (no visible way to redirect to AAD)
In summary, how do I set this setting with the new Authorization tool so that a MSI can make a cool to the function and authenticate with AAD.
...ANSWER
Answered 2021-Jun-08 at 08:10To make it work with the new Authentication
, follow the steps below.
1.Edit the Authentication settings
in the portal or set it when creating the app as below.
2.Edit the Identity provider
, make sure the Issuer URL
is https://sts.windows.net/
(without /v2.0
) and Allowed token audiences
include the App ID URI
.
For the App ID URI
, you could check it in your AD App of the function app -> Expose an API
, if you use the old Authentication
before, maybe it is your function app URL, it does matter, just make sure Allowed token audiences
include it.
3.Then in the datafactory web activity, also make sure the resource is the App ID URI
.
Then it will work fine.
Update:
You could refer to my configuration.
Function app:
AD App:
AD App manifest:
QUESTION
I want to add the timestamp of copying parquet files to my dataframe in data flow as a derived column.
In source module I can filter parquet files by last modified which makes me think that it should be possible to access files' metadata including copied timestamp through derived column transformations, but I couldn't find anything for it in Microsoft documentation.
ANSWER
Answered 2021-Jun-07 at 12:09There is no function can get the last modified time in the data flow expression.
As a workaround, you can create a Get Metadata activity to get that and then pass it's value to a parameter in your data flow.
The expression:@activity('Get Metadata1').output.lastModified
QUESTION
firstly I create a web activity to get keyvault,and then create a "set variable" activity. when I try to create variable in the "set variable" activity, it shows "no results found". BTW I cannot attach screenshot due to less reputation . I refer to this doc to do execution
...ANSWER
Answered 2021-Jun-03 at 09:07QUESTION
I am trying to use convertTimeZone to get to my local time in Sydney.
Data Factory is happy with with other conversion like
@convertTimeZone(utcnow() , 'UTC' , 'GMT Standard Time') but when I try for my location @convertTimeZone(utcnow() , 'UTC' , 'A.U.S. Eastern Standard Time')
I get an error
In the function 'convertTimeZone', the value provided for the time zone id 'A.U.S. Eastern Standard Time' was not valid.
it is in the list here https://docs.microsoft.com/en-us/previous-versions/windows/embedded/ms912391(v=winembedded.11)
Which is provided in the documentation here. https://docs.microsoft.com/en-us/azure/data-factory/control-flow-expression-language-functions#convertTimeZone
Any ideas ? Thanks
...ANSWER
Answered 2021-May-30 at 11:02Please see the description of destinationTimeZone in convertTimeZone function:
The name for the source time zone. For time zone names, see Microsoft Time Zone Index Values, but you might have to remove any punctuation from the time zone name.
So remove .
in A.U.S. Eastern Standard Time and then have a try this expression:
QUESTION
I'm using a column pattern to catch nulls. My logic is very simple.
Matching condition
...ANSWER
Answered 2021-May-20 at 20:51Try this instead - Decimal has precision and scale
startsWith(type, 'decimal')
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install data-factory
You can use data-factory like any standard Java library. Please include the the jar files in your classpath. You can also use any IDE and you can run and debug the data-factory component as you would do with any other Java program. Best practice is to use a build tool that supports dependency management such as Maven or Gradle. For Maven installation, please refer maven.apache.org. For Gradle installation, please refer gradle.org .
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page