How are containers partitioned?
Containers are partitioned horizontally across across compute within an Azure region and distributed across all Azure regions you configure in your Azure Cosmos DB for NOSQL account.
What types of throughput provisioning are provided by Azure Cosmos DB ?
- Container level throughput
- Database level throughput
- Mixed through put provisioning
State whether true or false:
- a container with provisioned throughput cannot be converted to shared database container. True
- A shared database container can be converted to have dedicated throughput. False
What is the currency in azure cosmos db sql?
Request units are a rate based currency.
What is the least amount of request units that can be provisioned ?
You cannot provision less than 400 RU/s and they are provisioned in increments of 100.
When migrating to Azure Cosmos DB for NoSQL what can help in estimating the cost in cosmos DB?
The Azure Cosmos DB Capacity Calculator is a calculator surfaced as an online form to plug in details about your existing data workload to help estimate your application storage and throughput requirements and translate it to a cost estimate in terms of azure cosmos db for No SQL.
What is TTL?
TTL is the time to live and is specified as integer in seconds.
How is TTL configured in a container?
TTL value is configured using DefaultTimeToLive property of the container.
What is the meaning of the following values for a TTL?
Does Not Exist – Items are not automatically purged
-1 – Items will not expire by default
n – n seconds after last modified time
How is TTL configured in a item?
The TTL value of an item is configured by setting the ttl path of the item. The TTL value will only work if the DefaultTimeToLive for the container is already set. If the TTL path for the item is set, it over rides the value of the DefaultTimeToLive.
What rate based currency acronym is used as a simplification of CPU, memory and IOPS?
RU/s
Which property of a container should be specified to automatically purge items after a specified number of seconds?
DefaultTimeToLive
What is azure cosmos db server less?
Azure cosmos db serverless is a consumption based model where each request consumes request units. The consumption model eliminates the need to pre-provision throughput requests ahead of time.
Specify the throughput model to use in the following cases –
- Which model will you use when global distribution is required? provisioned throughput. This is because server less can only run in a single azure region.
- Which model will you use in case there is a storage limit of 50 GB?serverless. This is because serverless only allows upto 50 gb, whereas with provisioned throughput we can store unliminted amount of data in the container.
What can be done in case we need to provide a range of throughput?
We should use autoscale for a provisioned throughput.
What is the minimum bill in case of autoscale throughput?
With autoscale, we can only set the maximum, and the minimum billed will be 10% of the maximum when there are 0 request.
In terms of RU/s when is autoscale vs provisioned used?
Throughput provisioning is ideal for scenarios where the full RU/s provisioned is consumed for more than 66% of hours per month. Autoscale throughput is helpful if your team cannot predict your throughput needs accurately or otherwise use the max throughput amount for less than 66% of hours per month.
What happens if the request exceed the max RU/s in provisioned throughput?
the upcoming requests are rate limited.
Is it possible to migrate to and fro autoscale ?
Yes
What is azure data factory?
Azure data factory is a native service to extract data, transform it and load it across sinks and stores in an entirely server less fashion.
How is cosmos db present within azure data factory?
Azure cosmos DB for NoSQL is available as a linked service within azure data factory. This linked service is supported both as source of data ingest and as a target of data output. For both, configuration is identical.
How is azure data factory configured when reading data from azure cosmos DB?
In azure data factory, when reading data from azure cosmos DB, we must configure our linked service as a source. To configure this, we must create a SQL query of the data we want to read in.
How is azure data factory configured when storing data into azure cosmos DB?
In azure cosmos DB, when storing data to azure cosmos DB for NOSQL, we must configure our linked service as a sink. To configure this, we must set our write behavior to either insert or upsert.
What is apache Kafka?
Apache Kafka is a open source platform used to stream events in a distributed manner.
How can data from Kafka move to another sources?
Kafka Connect is a tool to stream data within kafka and other data systems.
How can we connect Kafka to cosmos DB?
Four configuration properties should be set to properly configure connectivity to azure cosmos DB for NoSQL .
- connect.cosmos.connection.endpoint – Account endpoint URI
- connect.cosmos.master.key – account key
- connect.cosmos.databasename – name of the database source
- connect.cosmos.containers.topicname – using csv format, a mapping of the Kafka topics to containers.
Suppose you would like the products container to be mapped to the prodlistener topic and the customers container to the custlistener topic. What CSV mapping string should you use?
prodlistner#product,custlistner#customers
How can a topic be created in Kafka?
kafka-topics
How is a producer created in Kafka?
Kafka-console-producer
What is azure stream analytics?
Azure stream analytics is a real time event processing engine designed to process fast streaming data from multiple sources simultaneously. it can aggregate, analyse, transform and even move data around to other data sources for more profound and further analysis.
How are query results from stream analytics processed in azure cosmos DB?
Query results from azure stream analytics will be processed as a JSON output when written to azure cosmos DB for NoSQL. Items are inserted or upserted based on the id.
Which properties should be configured in azure cosmos DB for NOSQL when reading from stream analytics?
Output alias – an alias to refer to this output in the query
account Id – account endpoint URI
account key – acount key
database – name of the database resource
container name – name of the container.
What is the first step to take when connecting to azure synapse from cosmos?
Make sure that synapse link is enabled at the account level.
What should be done when creating a container with azure synapse enabled?
When creating a container, you should enable analytical storage at the container level on per container basis.
How can you read the analytical data of synpase from cosmos?
- You can choose to load to a spark dataframe where the meta data is cached.
- you can create a spark table that points to azure cosmos db for NoSQL directly.
Which type of component in azure data factory will load out to azure cosmos DB for Nosql after it has been transformed?
Sink
After enabling azure synapse link at the azure cosmos db for NoSQL account level, what should you do before you can use the spark container with a specific container?
enable analytical storage at the container level
Foolishly Yours,
Avantika Tanubhrt
Happy Learning 🙂

Leave a comment