Impact 2024: The Industrial Data and AI Conference for and by Users | Nominate Speakers Now for a Ch...
Is it possible to increase the limit on execution of workflow instances per project or limit on execution workflow instance itself instead of project ? as we require to schedule workflow, on that it depends how much data we get to run workflows instances, so it can be more than 50
I want to complete Data Engineer Basics - Integrate so, How to purchase 2 credits for Data Engineer Basics - Integrate?
Hello ! I am new to this community but would like to ask the following: Has anybody any experience extracting a list of DCS Process Alerts to Operators or Operator Overrides on an Hourly or Daily basis into CDF ? This data resides in the OT and is sometimes extracted and made available in reports in the IT domain, e.g., by Yokogawa. As I see it, this is an important indicator for stable and well managed production operations, and also a good predictor of upcoming threats to production.
Hello everyone! I have several binary timeseries that are displayed correctly when I ask for a short period of time: However, if I ask for a longer timespan, like a year, Charts gives me the aggregates, which I believe would be fine on some cases, but in this particular binary case, it does not make any sense. The final user cannot use this information to perform any analysis; any calculations, thresholds, or just “taking a look to see the trend” will not give an accurate result. Is it possible to choose to disable the use of aggregates for cases like this? Thank you in advance.
I’m working on a prototype for a flexible data model to store time series data in a way that is easy to catalogue, query and filter. Using Pygen both to populate and use the model seems convenient.At its current iteration, I’ve only applied direct relations and (undocumented?) @reverseDirectRelations in the GraphQL schema. I expected to be able do something similar to client.windmill(windfarm="Hornsea 1").blades(limit=-1).sensor_positions(limit=-1).query()as found in the Pygen documentation, but it does not work (my client.windmill analouge has no methods corresponding to its relations). Do I have to use edges instead of relations to query easily and declaratively with Pygen?
I’m trying to use Update extraction pipelines method to update my extraction pipelines created. Here are some issues i face:1. I’m unable to update contacts for an extraction pipeline, it throws the below error:code - contact_info = [ExtractionPipelineContact(name="sangs", email="sm6@slb.com", role = "MAINTAINER", send_notification=True)]to_update = ExtractionPipelineUpdate(external_id="<PIPELINE-EXT-ID-2>")to_update.contacts.set(contact_info)client.extraction_pipelines.update(to_update)error - “CogniteAPIError: Unexpected field - items[0].update.contacts.set[0].send_notification - did you mean sendNotification? | code: 400 | X-Request-ID: 918d1113-6a5e-9d38-b850-61e3dc54c220 The API Failed to process some items. Successful (2xx): [] Unknown (5xx): [] Failed (4xx): [<PIPELINE-EXT-ID-2>, ...]”2. I’m unable to set the specified fields to None using sdk : description, source, schedule,documentation, name and dataset-id. Im able to set only metadata and raw-tables to None. Wond
Hi, I’m doing the Learn to Use the Cognite Python SDK module in the Data Scientist basics seems incorrect.For example, I’m in the notebook `2_List_Search_Retrieve.ipynb` and the data presented appears to be very random, and not industrial data. When I run a query like:`c.assets.list(metadata={'ELC_STATUS_ID': '1211'},limit=5)` this comes back empty because there is no industrial data in any of the the datasets. In fact, almost all of the datsets in the c.data_sets.list() command are Deprecated and archived.It looks to me like potentially the datasets are tests that people have made. I would really like to be able to see some more realistic data in the dataset used for exploring the tool. Please let me know how I can do that, if anyone knows!
Hi Team,I am doing the Notebook 2 in Data Engineer Basics - Integrate course and getting this error, am I missing something here?
Hi Team,Please help me to find the mistakes in my code. Let me showcase the dummy codemain code:def main() -> None: """ Main entrypoint """ BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) config_values_vault.set_vara() with Extractor( name="SAP_Extractor", description="An extractor to extract asset hierarchy from SAP based on root node", config_class=SapConfig, # version=__version__, # debugger version="1.0.0", run_handle=run, metrics=metrics, config_file_path= os.path.join(BASE_DIR, 'config.yaml'), ) as extractor: extractor.run() I have build code unittest for above code: def test_main(): with patch('os.path') as path_mock: with patch.object(path_mock, 'abspath') as mock_abspath: with patch.object(path_mock, 'dirname') as mock_dirname: with patch.object(path_mock,'join') as mock_join: mock_abspath.return_value = '/p
Hello everyone.I have a use case that I have been trying to figure and I'm hoping you could help me.We have an equipment, let's say a pump, and we have a timeseries for its flow. We consider that the pump is operating as long as the flow is above a threshold.In PI, if I want to know how long the pump was operating, I can simply use PITimeFilterVal function in Excel to retrieve the amount of time that the timeseries was above the threshold. Me and my team have been trying to do this with Cognite, with no success. We tried transforming the timeseries to a 0 or 1 using the “Threshold” function, and then integrating it, but we have faced some limitations due to the approximations that are intrinsic to the integration function.Unfortunately, I cannot show you the actual data, be we checked a few days where the pump starts the day operating - at 00:00 - and at 2 am our integration gives us a value of 1.84h - when it should be as close as 2h as possible. This difference, even though slight, i
Data Workflow UI stopped rendering the workflow with the following error message. It stopped working as soon as I created a CDF function to trigger the workflow. https://delfi-us.fusion.cognite.com/shaya-dev/flows/SDM-ARP-Model-Refresh?cluster=az-eastus-1.cognitedata.com&env=az-eastus-1
Is there a way to query metadata of TimeSeries from a particular dataset in CDF Transformations?Something like this, but this query isn’t workingselect cast(`externalId` as STRING) as externalId, cast(`dataSetId` as STRING) as eidfrom `_cdf`.`timeseries`where dataSetId == <data_set_id>;
Facing Session IDP refresh from UI while running workflows and transformations. Tried using client id and client secret. Facing this since yesterday Let me know if any changes required from our end
Hi community,I have noticed a large discrepancy in runtime when running my handle function as a single Cognite Function call and when I run the handle function locally. And yes, this is the runtime of the actual CALL, not the creation/deployment of the Cognite Function. The handle function is doing some calculations on a time series with around 500 000 datapoints. Locally, it runs under 1 minute, while through Cognite Function call it takes around 9 minutes. Do you have any idea what could cause the large discrepancy?Thanks!
I am making comparisons between time series data in CDF and PI. The reason is that in our tenants the CDF data is not 100% accurate compared to PI.However, from my testing I think that PI performs its aggregations with the timestamps “centered” at the aggregated time periods, while CDF puts the timestamps at the start of each aggregated period. Is it possible to specify how this is done with the Python API? From my study of the docs it appears not to be the case. The same applies to the PI Web API as well: I cannot specify how the timestamps are placed. The agreement with PI becomes significantly better if I place the CDF timestamps at the center of the aggregated time periods.My current workaround is the following:Fetch RAW data from CDF Shift the timestamps by 0.5x of the granularity Resample to the desired granularity Compute mean Interplate any missing valuesThe issue is that fetching raw data is a lot more time consuming than fetching aggregates. I have been playing with fetching
Deployment of workflow , can you guide how we can automate deployment of cognite workflows using sdk
While testing versions of the cognite-sdk python >= 7.37.0, we noticed a performance issue in the "retrieve_dataframe" method of the time series API for all versions 7.37.x ( x >= 0 and x <= 3). Example: For the same data window, we tested with version 7.0.0 and the latest versions. The processing time for version 7.0.0 in this method was less than 5 seconds, while for newer versions the processing time was 3 minutes.
Hi Team,Will the data flow in Cognite is Uni-directional or Bi-directional? Thank you,Navyasri Indupalli
Hello team, I was trying to trigger a workflow by passing the workflow input tasks and setting the onFailure to skipTask, as I did not want my workflow to be aborted if some task fails.Despite of triggering the workflow by sdk and passing the workflow input.,I can see that the workflow has been trigerred but with onFailure to be abortWorkflowCould you please help me with solving this. code for triggering workflow:workflow_input = { "tasks": [ { "externalId": "gb-test-func-1", "type": "function", "name": "1710479974441", "parameters": { "function": { "externalId": "gb-test-func-1", "data": {} } }, "onFailure": "skipTask", "dependsOn": [] }, { "externalId": "gb-test-func-2", "type": "function", "name": "1710482066717", "parameters": { "function": { "externalId": "gb-test-func-2", "data": {} } }, "onFailure": "skipTask", "dependsOn": [ { "externalId": "gb-multiplication-by-2" }
The buttons have disappeared in the search after one of the releases. It is still usable, but you have to guess the functionalities.
Hi Team, Is there any possibility that I can register for Bootcamp virtually or it can be in India location as I am from India.Also, want to know is the Bootcamp is for individual or group and the cost for attending the Bootcamp? Thanks,Navyasri Indupalli
Hello, I am trying to perform a calculation on Charts but it does not work. It is a multiplication of 4 time series, all of them with a resample to granularity to 1m. One of them has real values and 3 of them are 0/1 steps. When I zoom out, I can see the calculation and it shows a pick of values, that is not accurate as it shows average data. Although when I zoon in to see the data, the trend disappears, the error says “One of the time series has less than two values”. As I resampled the data to 1 minute, I was expecting to see one values per minute, but it does not seems to happen. Is there other way to calculate the data correctly? Zoom in:
It was requested by @ibrahim.alsyed from Celanese that we increase the limit of the raw datapoints for the endpoint /timeseries/data/list (Retrieve data points) to 1 million. Currently, the non-aggregated data points returned is limited to 100000.For the drill-down views with more than 7 days, the server-to-client data transfer maximum size is reached due to the number of datapoints (some timeseries have more than 1 datapoint per minute). The solution they have implemented is that for ranges higher than 7 days, they are displaying an interpolated trend using the maximum absolute values for 30-minute aggregation. Once the user selects a smaller date range, they are unable to display all the values. Hence a increase in the limitation was requested.
I updated and published a new version of the flow. When I am trying to run the updated flow it is still running the older version.Below is the updated version - Below is the version after initiating the Run
Can Unified namespace architecture be implemented in Cognite Data Fusion
Already have an account? Login
Enter your E-mail address. We'll send you an e-mail with instructions to reset your password.
Sorry, we're still checking this file's contents to make sure it's safe to download. Please try again in a few minutes.
Sorry, our virus scanner detected that this file isn't safe to download.