Recently active
Hi all, my name is Mikey Martinez and I am joining the Houston team as a PM. I have been working in SaaS since 2016, first as a GIS analyst/PM for a small pipeline integrity software company permanently embedded at Exxon Mobil and then I made the switch to full time PM for a mid-size land data management software company. I have a background in Geology and have always found the oil and gas world fascinating and hope that with my diverse background I will be able to bring a lot of value to Cognite. Thanks!
We have an OPCUA extractor pulling data from an Ignition OPC Server which had been working well at a customer site for weeks. This is a try before you buy scenario so the issue of data loss notification is important and that we respond back to the customer as to reasons why this occurred. It seems something changed on the Ignition OPC server preventing the extractor from accessing the tag data. We can see that data stopped and when using the Cognite chart. However we did not receive an email notification of a pipeline failure over several days. Upon restarting the OPC Extractor service, we did receive the pipeline failure email as seen below. We know the data access problem is in the OPC Server so not trying to get help from Cognite with that but It seems that the pipeline SHOULD have sent a failure notification WITHOUT having to restart the service. That is the important fact to bring to Cognite’s attention. Failure email message: Error: Root node does not exist: ns=2;s=[IAWAT_Tags_En
Hi CDF teamThere is a tool to search for assets in pdf files that is available in "data management" but not available in "industrial tools", I find it very useful, but they should implement it in industrial tools, since it is the end user interface.
my team has uploaded all the laser scan (360) files, unit wise but some files are having some error I want to check file with file number but in 360 all file name is seen as unknown, how can I find the particular file
Hi, The user is listed as UNKNOWN when the annotation is done manually from “Data explorer”. This is problematic as we need to be able to track who made the changes. Here is an annotation I added manually: And this is what I see from the API. I am able to identify the creating_app which is data explorer, but I need to know the user as well or at least some way to track which credentials made the change. UNKNOWN is not good enough. I’ve also notices that for our automated annotations this field does not work as expected. These are all made by the same user/credentials, but for some reason the are all different and it looks like the “creating_user” field is mapped to the “job_id” you get when sending a request to the API rather than the user that made the request. I don’t know about others, but we at least need to be able to track which users/credentials created the annotations, which batch job uploaded the annotation is not relevant to us at all.job.7445037198441218job.5162020989972303j
I have created 1 workflow , in which I am creating dynamic tasks depending on input, it creates batch of ids and create tasks out of it. Below is workflow definition WorkflowVersionUpsert( workflow_external_id="test_dynamic-0729", version="1", workflow_definition=WorkflowDefinitionUpsert( description="This workflow has two steps", tasks=[ WorkflowTask( external_id="test_sub_tasks", parameters=FunctionTaskParameters( external_id="test_sub_tasks", data="${workflow.input}" ), retries=1, timeout=3600, depends_on=[], on_failure = "abortWorkflow", ), WorkflowTask( external_id="test_create_sub", parameters=DynamicTaskParameters( tasks="${test_sub_tasks.output.response.tasks}" ), name="Dynamic Task", description="Executes a list of workflow tasks for subscription creation", retries=0, timeout=3600, depends_on=["test_sub_tasks"], on_failure = "abortWorkflow", ) ] ) As part of this workflow, I have some task that are needed to be executed in parallel and are expected to finish in around similar
When using time series data, there are often situations where you want to get point in time for each day.For example, for time series A, which records the accumulated value, the raw data at 9 o'clock on 1/1, the raw data at 9 o'clock on 1/2, and so on.Even if you specify "Granularity=1d" using the SDK, the data obtained is aggregated, so there is a discrepancy from the raw data.Therefore, when I acquire raw data with daily granularity, I pick up the raw data as of 9 o'clock on 1/1 after acquiring raw data with fine granularity.This is very tedious and time-consuming, so I want an easy way to get the raw data at a specific time from the SDK options or UI.
When I open my Canvas, I have to manually enable auto refresh on each item. For a dashboard that a user might use every day, this is not feasible. Can there be a way to enable auto refresh for all items on a Canvas simultaneously?
Problem Statement: Synthetic Time Series in CDF are dynamically calculated based on expressions, but lack persistent identifiers. This limits usability when users need: Persistent IDs for discovery and access. Simplified queries via API, SDK. Centralized definition of expressions as part of the data model, avoiding the need for each backend to redefine them. Suggested Approach: Introduce a Synthetic Time Series Definition object that: Allows defining synthetic series with a persistent external_id. Stores metadata like expressions, descriptions, and units. Enables dynamic evaluation without requiring data storage. Supports defining expressions as part of a model, enabling reuse across different systems without requiring redundant definitions in backends. Benefits: Usability: Persistent identifiers for easier access and queries. Consistency: Eliminates repetitive expression definitions. Scalability: Centralized expression definitions simplify updates and maintenance.
We’re constantly innovating to bring you engaging ways to enhance your skills and showcase your expertise. As part of this commitment, we’re introducing Microlearning Bundle Badges —an exciting way to reward your dedication to learning!✨ These badges recognize your completion of curated course bundles, designed to provide a deeper understanding of key topics and tools. They’re perfect for learners looking to consolidate their knowledge and demonstrate their proficiency. We’ve rolled out the following badges: Microlearning Bundle badge- Canvas Microlearning Bundle badge- Data Modeling Microlearning Bundle badge- Data Workflows This is just the beginning—more bundles and badges are on the horizon!🚀 Got ideas? We want to hear from you! If there’s a specific topic or skill you’d love to see in a future microlearning offering, share your ideas in the Product Ideas section under the Academy Trainings tab. Your feedback helps us create the best learning experience possible! Start earning you
Hi, On behalf of Celanese, I’m working on creating Data Workflows for a project that run transformations requiring specific credentials. The credentials set in the Workflow Trigger can call the transformations, but they do not (and should not) have the same level of permissions as the credentials defined for the transformations themselves. The issue I’m facing is that transformations are being executed in “current user” mode instead of “client credentials” mode which makes the transformation throw a missing capability error. Unfortunately, there’s no option to specify this behavior when setting up a Workflow Task. Is there a way to overcome this limitation?
Hello team, We have tried to query Entity view which has a property properties of type view Property which is a reverse direct relation. We are using the query endpoint and the instances.query() sdk method to do so. We want the details of properties to come in the select object of Entity. Providing the query below: { "with": { "0": { "limit": 50, "nodes": { "filter": { "and": [ { "matchAll": {} }, { "hasData": [ { "type": "view", "space": "slb-pdm-dm-governed", "externalId": "Entity", "version": "1_7" } ] }, { "or": [ { "equals": { "property": [ "node", "externalId" ], "value": "fba80a5d3b994db698e74b77fb96f1de" } } ] } ] } } }, "0_6": { "limit": 10000, "nodes": { "from": "0", "through": { "source": { "type": "view", "space": "slb-pdm-dm-governed", "externalId": "Property", "version": "1_7" }, "identifier": "entity" }, "direction": "inwards", "filter": { "and": [ { "matchAll": {} }, { "hasData": [ { "type": "view", "space": "slb-pdm-dm-governed", "externalId": "Property", "version": "1
As an improvement to Cognite Functions. It would be nicer to have the data entered as parameters instead of writing the entire JSON. Internally, CDF can generate the appropriate JSON as data dictionary When invoking the function, you get the list of mandatory parameters, types of each (string, number, array, object...etc), description as well as the default value for each. When having an array parameter for example, you can add elements to the array or remove elements...etc. Essentially, it is a glorified JSON builder UI This allows for better user experience as you do not have to remember what parameters the function requires, what is mandatory or not, what the names and types of parameters are and if not set, what are the default values.
Hi! It would be great to have a possibility to add / change the placeholder text for Cognite functions. Example: It would be great to be able to add placeholder from code.
When making calls to the Cognite API, successful calls provide no confirmation of success. HTTP Response 200 would be helpful for asserting successful API requests. One point of uncertainty for us is when passing multiple External ID’s to delete row records in a staging table, if some ID’s are incorrectly identified, we aren’t receiving any feedback to indicate 100% success. We’ve read that when passing batches, there should be a return item indicating which records failed, but that is not our experience. Note: this is using the Cognite SDK.
Hi For a dashboard use case we are working on I want to extract a list of the column names in each raw table we have in our staging area. At the moment, it does not seem to be a way of doing this. I have made two very hacky ways of accessing this information (see the code example below), but they are either very time consuming because of inferring the raw schema, or it does not return anything, because the table has two many columns and it times out. This makes this method unfeasible when running the scripts for our whole environment, which would need to happen regularly. I feel like there has to be a better way of doing this. I know raw is a schemaless service, but the columns do exist. Having this information would greatly improve our efforts in getting a better overview of our data. from pydantic import BaseModel, Field class RawTable(BaseModel): database: str table: str def to_friendly_name(self) -> str: return f"{self.database}.{self.table}" def get_inferred_raw_schema(self, cogni
We are doing quite a few transformations on larger FDM-tables, and are experiencing them to be a lot slower than what is to be expected. From my understanding CDF is reading some rows a at a time from the database, doing a standard database operation (with Spark), and writing back to the database. The read operation is very slow (for one transform I am running now read speed is 1600 rows/s). Is it possible to write a transformation (with standard SQL) that will be run in-database, that avoids the slow read? This would be especially beneficial for FDM to FDM transformations, as this would reduce time from minutes to milliseconds for some of our transformations.
When saving & scheduling a calculation, how do I specify the offset? For example, I can specify that I want the calculation to run every day, but how do I specify that I want it to run at 12pm every day?
Is there a recommended approach for using YAML configurations to automate the creation of spaces, containers, views, and data models in Cognite Data Fusion? How can we incorporate customizable parameters (e.g., space names, descriptions, and container properties) in the YAML files to make the process more flexible? Example YAML configuration for containers: containers: - name: "example_container_1" description: "First sample container" external_id: "example_container_1_id" properties: name: type: "Text" nullable: false parent: type: "DirectRelation" nullable: true isValid: type: "Boolean" nullable: true indexes: - index_name: "entity_name" type: "BTree" properties: ["name"] - name: "example_container_2" description: "Second sample container" external_id: "example_container_2_id" properties: identifier: type: "Text" nullable: false indexes: - index_name: "identifier_index" type: "BTree" properties: ["identifier"] Each container is defined with the following: Properties: Each property ha
Do you have the ability to hide left hand navigation items in the different workspaces? For example if your organization isn’t using Customer Apps (BETA) can you hide that from the menu options?
Welcome to the CDF Fundamentals Discussion! This discussion is dedicated to help learners of the Cognite Data Fusion Fundamentals learning path succeed. If you’re struggling with the exercises in this learning path, try the tips & tricks below or post a comment with the challenge you’re facing. You can also post your own tips and respond to fellow learners’ questions. Cognite Academy’s instructors are also here to help.
Hi, I am on the CDF Fundamentals learning path and is currently on the “Events Transformation” section. I am trying to create a data set named “FirstnameBirthyear-IFSDB”. I do not see an attached dataset at the bottom of the page, do I use the same dataset (assets CSV file) as before? The external ID for the new IFSDB dataset is 8667716811777202. Thanks!
Hi, I’m going through the steps for the course “CDF Fundamentals”, and stuck on Working with CDF: Integrate → Asset transformation → Switch to SQL editor (enter code). Initially I was able to enter the code with no problem, and replaced all my data “Stacey1994”.. in all the applicable places. When the error occurred for the first time, I refreshed the page and tried to repeat the steps, but when I got to the “For NULL values on updates: Keep existing values (Default)”, the error message below keeps popping up. Thanks, Stacey
Hello everyone, I am Munjal , Mechanical Engineer with 17+ years of experience providing values in domains of Master Data Management, CMMS, Digital Transformations, Digital Twin, ERP systems such as SAP-PM/MM & Maximo, Mechanical Integrity for the clients in Oil and Gas, Petrochemical, Medical and CPG sectors.
Hi Team, I'm facing a problem when doing two-step authentication at: https://fusion.cognite.com I'm using cognite-learn as a domain name, using my registered Cognite account. However, when I try to authenticate, I choose my account, a verification code (two-digit number) for approval via Microsoft Authenticator is generated but it's not being sent to my mobile device Microsoft Authenticator App. I've tried other authentication options, cleared the browser cache and cookies, restarted the mobile device where MS Authenticator is located and checked if the app is updated to the latest version. MS Authenticator is working correctly with the other accounts I have and I didn't have this problem when I was doing the CDF Fundamentals training steps. Could someone please help with this?
Already have an account? Login
Enter your E-mail address. We'll send you an e-mail with instructions to reset your password.
Sorry, we're still checking this file's contents to make sure it's safe to download. Please try again in a few minutes.
Sorry, our virus scanner detected that this file isn't safe to download.
We use cookies to enhance and personalize your experience. If you accept you agree to our full cookie policy. Learn more about our cookies.
We use 3 different kinds of cookies. You can choose which cookies you want to accept. We need basic cookies to make this site work, therefore these are the minimum you can select. Learn more about our cookies.