会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 2. 发明授权
    • Data processing system, processor and method of data processing in which local memory access requests are serviced on a fixed schedule
    • 数据处理系统,处理器和数据处理方法,其中本地存储器访问请求以固定的时间表被服务
    • US07447844B2
    • 2008-11-04
    • US11457322
    • 2006-07-13
    • Leo J. ClarkGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Leo J. ClarkGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/1425G06F12/0817G06F12/0897
    • A processing unit includes a local processor core and a cache memory coupled to the local processor core. The cache memory includes a data array, a directory of contents of the data array. The cache memory further includes one or more state machines that service a first set of memory access requests, an arbiter that directs servicing of a second set of memory access requests by reference to the data array and the directory on a fixed schedule, address collision logic that protects memory access requests in the second set by detecting and signaling address conflicts between active memory access requests in the second set and subsequent memory access requests, and dispatch logic coupled to the address collision logic. The dispatch logic dispatches memory access requests in the first set to the one or more state machines for servicing and signals the arbiter to direct servicing of memory access requests in the second set according to the fixed schedule.
    • 处理单元包括本地处理器核心和耦合到本地处理器核心的高速缓存存储器。 高速缓冲存储器包括数据阵列,数据阵列的内容目录。 缓存存储器还包括服务于第一组存储器访问请求的一个或多个状态机,通过参考数据阵列和固定时间表上的目录来指导第二组存储器访问请求的服务的仲裁器,地址冲突逻辑 其通过检测和发出第二组中的活动存储器访问请求与后续存储器访问请求之间的地址冲突以及耦合到地址冲突逻辑的调度逻辑来保护第二组中的存储器访问请求。 调度逻辑将第一组中的存储器访问请求分派到一个或多个状态机用于服务,并且向仲裁器发出信号,以根据固定的时间表对第二组中的存储器访问请求进行直接服务。
    • 4. 发明授权
    • Interlock for controlling processor ownership of pipelined data for a
store in cache
    • 用于控制缓存中存储的流水线数据的处理器所有权的联锁
    • US5490261A
    • 1996-02-06
    • US680176
    • 1991-04-03
    • Bradford M. BeanAnne E. BierceNeal T. ChristensenLeo J. ClarkSteven T. ComfortChristine C. JonesPak-Kin Mak
    • Bradford M. BeanAnne E. BierceNeal T. ChristensenLeo J. ClarkSteven T. ComfortChristine C. JonesPak-Kin Mak
    • G06F9/38G06F12/08G06F12/00
    • G06F12/0811
    • Insures data integrity in process ownership indications by providing an ownership interlock on the data units in a pipeline to a store-in type of cache. An ownership interlock prevents any processor ownership change to occur (i.e. exclusive or readonly ownership) for a cache data unit until all outstanding stores have been made in the cache data unit, after which the ownership may be changed. An ownership change may be signalled by a cross-invalidate (XI) signal to a processor. Outstanding stores are received by the pipeline after the stores are completed by a processor, and the outstanding stores output from the pipeline into a store-in cache. A continuous flow of stores is enabled into and out of the pipeline to expedite a change of ownership requested of a data unit in the cache. The continuous flow avoids having to stop a processor from putting stores into the pipeline and avoids forcing all outstanding stores out of the pipeline into the cache before indicating a change of processor ownership.
    • 通过在流水线中的数据单元上提供所有权互锁到存储型缓存来保护进程所有权指示中的数据完整性。 所有权互锁防止对高速缓存数据单元发生任何处理器所有权改变(即,独占或只读所有权),直到所有未完成的存储已经在高速缓存数据单元中进行,之后可以改变所有权。 所有权变更可以通过交叉无效(XI)信号发送给处理器。 在存储由处理器完成之后,流水线接收到未完成的存储,并且从流水线输出的未完成存储到存储缓存中。 连续的商店流程被启用进出管道,以加快对高速缓存中数据单元所需的所有权的更改。 连续流程避免了停止处理器将存储放入流水线中,并避免在指示处理器所有权的更改之前将所有未完成的存储从管道中强制进入高速缓存。
    • 5. 发明授权
    • Protecting ownership transfer with non-uniform protection windows
    • 用不均匀的保护窗保护所有权转让
    • US08205024B2
    • 2012-06-19
    • US11560619
    • 2006-11-16
    • Leo J. ClarkJames S. Fields, Jr.Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Leo J. ClarkJames S. Fields, Jr.Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F3/00G06F13/00
    • G06F15/173
    • In a data processing system, a plurality of agents communicate operations therebetween. Each operation includes a request and a combined response representing a system-wide response to the request. Latencies of requests and combined responses between the plurality of agents are observed. Each of the plurality of agents is configured with a respective duration of a protection window extension by reference to the observed latencies. Each protection window extension is a period following receipt of a combined response during winch an associated one of the plurality of agents protects transfer of coherency ownership of a data granule between agents. The plurality of agents employing protection window extensions in accordance with the configuration, and at least two of the agents have protection window extensions of differing durations.
    • 在数据处理系统中,多个代理之间进行通信。 每个操作包括一个请求和组合的响应,代表对该请求的全系统响应。 观察到请求的延迟和多个代理之间的组合响应。 通过参考所观察到的延迟,多个代理中的每个被配置有保护窗口扩展的相应持续时间。 每个保护窗口扩展是在绞盘期间接收到组合响应之后的周期,多个代理之一相关联的一个代理保护代理之间的数据粒子的一致性所有权的传送。 多个代理根据配置​​使用保护窗口扩展,并且至少两个代理具有不同持续时间的保护窗口扩展。
    • 9. 发明申请
    • Data Processing System, Processor and Method of Data Processing in which Local Memory Access Requests are Serviced on a Fixed Schedule
    • 数据处理系统,处理器和数据处理方法,其中本地存储器访问请求在固定时间表上服务
    • US20080016278A1
    • 2008-01-17
    • US11457322
    • 2006-07-13
    • Leo J. ClarkGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Leo J. ClarkGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/1425G06F12/0817G06F12/0897
    • A processing unit includes a local processor core and a cache memory coupled to the local processor core. The cache memory includes a data array, a directory of contents of the data array. The cache memory further includes one or more state machines that service a first set of memory access requests, an arbiter that directs servicing of a second set of memory access requests by reference to the data array and the directory on a fixed schedule, address collision logic that protects memory access requests in the second set by detecting and signaling address conflicts between active memory access requests in the second set and subsequent memory access requests, and dispatch logic coupled to the address collision logic. The dispatch logic dispatches memory access requests in the first set to the one or more state machines for servicing and signals the arbiter to direct servicing of memory access requests in the second set according to the fixed schedule.
    • 处理单元包括本地处理器核心和耦合到本地处理器核心的高速缓存存储器。 高速缓冲存储器包括数据阵列,数据阵列的内容目录。 缓存存储器还包括服务于第一组存储器访问请求的一个或多个状态机,通过参考数据阵列和固定时间表上的目录来指导第二组存储器访问请求的服务的仲裁器,地址冲突逻辑 其通过检测和发出第二组中的活动存储器访问请求与后续存储器访问请求之间的地址冲突以及耦合到地址冲突逻辑的调度逻辑来保护第二组中的存储器访问请求。 调度逻辑将第一组中的存储器访问请求分派到一个或多个状态机用于服务,并且向仲裁器发出信号,以根据固定的时间表对第二组中的存储器访问请求进行直接服务。
    • 10. 发明授权
    • Data processing system, method and interconnect fabric having a flow governor
    • 具有流量调节器的数据处理系统,方法和互连结构
    • US08254411B2
    • 2012-08-28
    • US11055399
    • 2005-02-10
    • Leo J. ClarkGuy L. GuthrieWilliam J. Starke
    • Leo J. ClarkGuy L. GuthrieWilliam J. Starke
    • H04J3/16H04J3/22G06F13/00G06F15/173
    • H04L47/722H04L47/70H04L49/109
    • A data processing system includes a plurality of local hubs each coupled to a remote hub by a respective one a plurality of point-to-point communication links. Each of the plurality of local hubs queues requests for access to memory blocks for transmission on a respective one of the point-to-point communication links to a shared resource in the remote hub. Each of the plurality of local hubs transmits requests to the remote hub utilizing only a fractional portion of a bandwidth of its respective point-to-point communication link. The fractional portion that is utilized is determined by an allocation policy based at least in part upon a number of the plurality of local hubs and a number of processing units represented by each of the plurality of local hubs. The allocation policy prevents overruns of the shared resource.
    • 数据处理系统包括多个本地集线器,每个集线器通过相应的一个多个点对点通信链路耦合到远程集线器。 多个本地集线器中的每一个排队对存储器块进行访问的请求,用于在到远程集线器中的共享资源的点对点通信链路中的相应一个上传输。 多个本地集线器中的每一个仅利用其相应点对点通信链路的带宽的小数部分向远程集线器发送请求。 所使用的分数部分由至少部分地基于多个本地集线器的数量和由多个本地集线器中的每一个表示的多个处理单元的分配策略确定。 分配策略可以防止超出共享资源。