不幸的是,这些技术往往都涉及到打破块级别上的数据并重新分配,或根据元数据表格将其删除。
Unfortunately these techniques often involve breaking up data at the block level and redistributing it, or deleting it based on meta-data tables.
虽然这意味着数据永远不会写入到位(从来没有被覆盖),而是写入新块并更新元数据来引用数据。
This means that data is never written in place (never overwritten), but instead new blocks are written and the metadata updated to reference it.
网络不习惯于用较大的数据块或较小的字节来表示数据,而习惯于以包为单位。
Networks tend to consider data not in large chunks, nor in tiny bytes, but in packets.
当表示客户的每个数据块到达本地计算机以进行处理时,数据块将临时放在临时目录中。
As each block of data representing a customer arrives in the local machine for processing, it is placed temporarily in a staging directory.
处理器内核随后选出未压缩的、过滤后的数据块,执行基本的数据库操作,如排序、联接和聚合。
A processor core then picks up the uncompressed, filtered data block and performs fundamental database operations such as sorts, joins, and aggregations.
数据对象类型表示原语的分组和其他数据对象类型,是数据图的构建块。
DataObjectTypes represent groupings of primitives and other DataObjectTypes, and are the building blocks of data graphs.
顾名思义,内存中数据库完全驻留在主存中,而不是磁盘上,这使得数据访问比传统的基于磁盘的数据块要快一个数量级。
As the name implies, an in-memory database resides entirely in main memory rather than on disk, making data access an order of magnitude faster than with conventional, disk-based databases.
尽管缓存可能在一定程度上减轻这种性能影响,但如果使用特定大小的数据块检索数据来支持用户界面分页,则可能出现另一个问题。
Although caching might alleviate this to some degree, another problem might arise if data is retrieved in blocks of a certain size to support user interface pagination.
处理器可能会在将一个数据块交还给内存前对这块数据执行数十或数百次操作。
The processor might perform dozens or hundreds of operations on a chunk of data in the cache before relinquishing it to memory.
此脚本可以是在我们将数据块传递给管道进行数据摄入之前需要的任何内容。
This script can be anything we want before the chunk is passed to the pipelines for ingestion.
使用与关系数据库一起提供的块载入工具来初始化数据库表。
Use the bulk load tools that are provided with relational databases to initialize database tables.
数据根据节点分块,所以这些查询能分别在各自的数据块,通过一个索引过的范围查询得到结果。
Data is partitioned by node, so these queries can each be answered by a single partition, using an indexed range query.
随后我将初始向量(IV)抽取到字节数组,对明文数据进行块加密,并将iv字节置于加密的数据字节之前。
I then extract the initialization vector (IV) into a byte array, block encrypt the clear data, and prefix the IV bytes to the encrypted data bytes.
当然,这不能彻底解决不及时写数据块的问题,但是,更频繁地写元数据也确实促进了更频繁地写数据。
Of course, this does not directly address the problem of not writing data blocks in time, but writing metadata more frequently does encourage data to be written more frequently as well.
没有对原始数据的映射,现在只有新的数据块在消耗文件系统中的磁盘空间。
Instead of a map to the original data, there are now new blocks with data, consuming disk space in the file system.
不仅不可能同时将多个数据块写入磁盘,而且每当进行修改时马上将每个数据块写入磁盘,也对系统性能有不利影响。
Not only is it impossible to write multiple data blocks to disk simultaneously, but writing every data block to disk when any part of it changes would be bad for system performance.
按大小拆分功能与按分隔符拆分类似,它也是将文件分为多个较小的数据块,然后再将这些数据块逐个传输到端点组件。
The split-by-size feature is similar to splitting by delimiter because you use it to divide a file into smaller chunks and transfer them to the endpoint component one by one.
应该经常这样做:在数据从数据库出来的过程中转换数据时,将较小的数据块放到网络上。
And you'll often be doing exactly that: Putting smaller chunks of data on the network, when you transform the data on its way out of the database.
CDATA(字符数据)区域被解析器视为数据块,从而允许您在数据流中包含任意字符。
CDATA (character data) sections are treated as a block of data by the parser, allowing you to include any character in the data stream.
数据库缓冲池通常是数据库共享内存中最大的一块内存。
The database buffer pool(s) area is normally the largest component of the database shared memory.
缓冲区缓存——为存储从物理数据文件读取的数据块而分配的内存。
Buffer Cache - Memory allocated for storing data blocks read from physical data files.
出现这种情况的原因如下:元数据更新删除了对原先数据块的任何引用,却指向磁盘上没有填充过数据的数据块。
Here's why this could happen: the metadata update would erase any reference to the original data block(s), but would point to data block(s) on disk that were never filled with data.
当从加载数据块队列取出一个数据块时,将执行该数据块的前脚本。
When a chunk is taken from the load chunk queue, the chunk pre-script is executed.
紧凑的数据表可以针对基于行的数据库做存储块优化,让SQL执行得更好。
Compact tables can optimize storage blocks for row-based databases and accordingly make SQL perform better.
如果散列值不相同,将过期的数据块标记为不同步,随后的同步确保数据块是正确同步的。
But if the hashes differ, the out-of-date block is marked as out of sync, and subsequent synchronization ensures that the block is properly synchronized.
牢记数据库的构建块及其继承关系:数据库由表组成,表由字段组成。
Always keep in mind the building blocks of databases and their hierarchical relationships: databases consist of tables, which consist of fields.
在将数据块真正发送给管道进行摄入之前,将执行该数据块的前脚本。
Before the chunk is actually sent to the pipelines for ingestion, the pre-script for the chunk is executed.
每次通过管道处理数据块时,会执行该数据块的后脚本。
Every time a chunk is processed by the pipelines, the post-script for the chunk is executed.
如果应用程序发送的数据大于32K,那么应用程序必须将大型数据分割成32k的数据块,并进行多次dpl。
If an application had to send more than 32k of data, then the application would handle it by chunking the large data into 32k blocks and doing multiple DPLs.
数据块:umf文件的一部分,数据块用于通过管道逐步处理UMF文件。
Chunk: a part of a UMF file, chunks are used to gradually process the UMF files by the pipelines.
应用推荐