Senad S Durakovic, Age 61410 Sheridan Ave APT 222, Palo Alto, CA 94306

Senad Durakovic Phones & Addresses

410 Sheridan Ave APT 222, Palo Alto, CA 94306 (781) 530-7460

193 River St, Waltham, MA 02453 (781) 647-5358

193 River St #B, Waltham, MA 02453 (781) 647-5358

6307 Bluff Springs Rd, Austin, TX 78744 (512) 707-9759

30 12Th St, Indiana, PA 15701 (724) 349-4197

Mentions for Senad S Durakovic

Senad Durakovic resumes & CV records

Resumes

Senad Durakovic Photo 20

Principal Engineer - Ai And Ml Modeling*> Compiler

Location:
Palo Alto, CA
Industry:
Semiconductors
Work:
Cavium Inc Apr 2018 - Jul 2018
Principal Engineer
Marvell Semiconductor Apr 2018 - Jul 2018
Principal Engineer - Ai and Ml Modeling*> Compiler
Amd Apr 2014 - Jan 2018
Gpu Architecture and Modeling Group Manager
Intel Corporation Feb 2001 - Apr 2014
Senior Component Design Engineer
Amd 1999 - Feb 2001
Rtl Design Engineer
Biocontrol Technology 1997 - 1999
Dsp Hw Engineer
Spectrum 1994 - 1997
Dsp Fpga and Board Development
Eka Elektronik Kontrol Aletleri Jun 1992 - Oct 1994
Arastirma Muhendisi
Bentas Jun 1991 - Aug 1992
Elektronik Muhendisi
Step5 Apr 1989 - Jun 1991
Jedan Od Osnivaca Kompanije
Education:
University of Sarajevo 1982 - 1987
Bachelors, Bachelor of Science
Skills:
Systemverilog, Asic, Computer Architecture, Verilog, Logic Design, Processors, C++, Embedded Systems, Soc, Rtl Design, Digital Signal Processors, Debugging, C, Semiconductors, Low Power Design, Pcb Design, Analog Circuit Design, Fpga
Languages:
Bosnian
Turkish
Senad Durakovic Photo 21

Senad Durakovic

Location:
Greater Boston Area
Industry:
Semiconductors

Publications & IP owners

Us Patents

Architecture To Support Color Scheme-Based Synchronization For Machine Learning

US Patent:
2021024, Aug 5, 2021
Filed:
Apr 22, 2021
Appl. No.:
17/237752
Inventors:
- Singapore, SG
Senad DURAKOVIC - Palo Alto CA, US
Gopal NALAMALAPU - Santa Clara CA, US
International Classification:
G06F 9/48
G06F 3/06
G06N 20/00
G06F 9/52
Abstract:
A system to support a machine learning (ML) operation comprises an array-based inference engine comprising a plurality of processing tiles each comprising at least one or more of an on-chip memory (OCM) configured to maintain data for local access by components in the processing tile and one or more processing units configured to perform one or more computation tasks on the data in the OCM by executing a set of task instructions. The system also comprises a data streaming engine configured to stream data between a memory and the OCMs and an instruction streaming engine configured to distribute said set of task instructions to the corresponding processing tiles to control their operations and to synchronize said set of task instructions to be executed by each processing tile, respectively, to wait current certain task at each processing tile to finish before starting a new one.

Array-Based Inference Engine For Machine Learning

US Patent:
2021005, Feb 25, 2021
Filed:
Oct 2, 2020
Appl. No.:
16/948867
Inventors:
- Singapore, SG
Ulf Hanebutte - Gig Harbor WA, US
Senad Durakovic - Palo Alto CA, US
Hamid Reza Ghasemi - Sunnyvale CA, US
Chia-Hsin Chen - Santa Clara CA, US
International Classification:
G06F 9/38
G06N 20/00
G06N 20/10
G06F 17/16
G06F 15/78
Abstract:
An array-based inference engine includes a plurality of processing tiles arranged in a two-dimensional array of a plurality of rows and a plurality of columns. Each processing tile comprises at least one or more of an on-chip memory (OCM) configured to load and maintain data from the input data stream for local access by components in the processing tile and further configured to maintain and output result of the ML operation performed by the processing tile as an output data stream. The array includes a first processing unit (POD) configured to perform a dense and/or regular computation task of the ML operation on the data in the OCM. The array also includes a second processing unit/element (PE) configured to perform a sparse and/or irregular computation task of the ML operation on the data in the OCM and/or from the POD.

Single Instruction Set Architecture (Isa) Format For Multiple Isas In Machine Learning Inference Engine

US Patent:
2019024, Aug 8, 2019
Filed:
Dec 19, 2018
Appl. No.:
16/226508
Inventors:
- Santa Clara CA, US
Ulf HANEBUTTE - Gig Harbor WA, US
Senad DURAKOVIC - Palo Alto CA, US
Hamid Reza GHASEMI - Sunnyvale CA, US
Chia-Hsin CHEN - Santa Clara CA, US
International Classification:
G06F 9/38
G06N 5/04
G06N 20/00
Abstract:
A programmable hardware system for machine learning (ML) includes a core and an inference engine. The core receives commands from a host. The commands are in a first instruction set architecture (ISA) format. The core divides the commands into a first set for performance-critical operations, in the first ISA format, and a second set of performance non-critical operations, in the first ISA format. The core executes the second set to perform the performance non-critical operations of the ML operations and streams the first set to inference engine. The inference engine generates a stream of the first set of commands in a second ISA format based on the first set of commands in the first ISA format. The first set of commands in the second ISA format programs components within the inference engine to execute the ML operations to infer data.

Array-Based Inference Engine For Machine Learning

US Patent:
2019024, Aug 8, 2019
Filed:
Dec 19, 2018
Appl. No.:
16/226539
Inventors:
- Santa Clara CA, US
Ulf HANEBUTTE - Gig Harbor WA, US
Senad DURAKOVIC - Palo Alto CA, US
Hamid Reza GHASEMI - Sunnyvale CA, US
Chia-Hsin CHEN - Santa Clara CA, US
International Classification:
G06F 15/80
G06N 5/04
G06N 20/00
G06F 15/78
G06F 9/38
Abstract:
An array-based inference engine includes a plurality of processing tiles arranged in a two-dimensional array of a plurality of rows and a plurality of columns. Each processing tile comprises at least one or more of an on-chip memory (OCM) configured to load and maintain data from the input data stream for local access by components in the processing tile and further configured to maintain and output result of the ML operation performed by the processing tile as an output data stream. The array includes a first processing unit (POD) configured to perform a dense and/or regular computation task of the ML operation on the data in the OCM. The array also includes a second processing unit/element (PE) configured to perform a sparse and/or irregular computation task of the ML operation on the data in the OCM and/or from the POD.

Architecture For Irregular Operations In Machine Learning Infference Engine

US Patent:
2019024, Aug 8, 2019
Filed:
Dec 19, 2018
Appl. No.:
16/226559
Inventors:
- Santa Clara CA, US
Ulf HANEBUTTE - Gig Harbor WA, US
Senad DURAKOVIC - Palo Alto CA, US
Hamid Reza GHASEMI - Sunnyvale CA, US
Chia-Hsin CHEN - Santa Clara CA, US
Rishan TAN - Santa Clara CA, US
International Classification:
G06F 17/16
G06F 15/78
G06N 20/00
G06N 5/04
Abstract:
A processing unit of an inference engine for machine learning (ML) includes a first data load steamer, a second data load streamer, an operator component, and a store streamer. The first data load streamer streams a first data stream from an on-chip memory (OCM) to the operator component. The second data load streamer streams a second data stream from the OCM to the operator component. The operator component performs a matrix operation on the first data stream and the second data stream. The store streamer receives a data output stream from the operator component and to store the data output stream in a buffer.

Streaming Engine For Machine Learning Architecture

US Patent:
2019024, Aug 8, 2019
Filed:
Dec 19, 2018
Appl. No.:
16/226534
Inventors:
- Santa Clara CA, US
Ulf HANEBUTTE - Gig Harbor WA, US
Senad DURAKOVIC - Palo Alto CA, US
Hamid Reza GHASEMI - Sunnyvale CA, US
Chia-Hsin CHEN - Santa Clara CA, US
International Classification:
G06N 5/04
G06N 20/10
G06N 3/063
G06F 9/54
G06F 12/06
Abstract:
A programmable hardware system for machine learning (ML) includes a core and a streaming engine. The core receives a plurality of commands and a plurality of data from a host to be analyzed and inferred via machine learning. The core transmits a first subset of commands of the plurality of commands that is performance-critical operations and associated data thereof of the plurality of data for efficient processing thereof. The first subset of commands and the associated data are passed through via a function call. The streaming engine is coupled to the core and receives the first subset of commands and the associated data from the core. The streaming engine streams a second subset of commands of the first subset of commands and its associated data to an inference engine by executing a single instruction.

Architecture Of Crossbar Of Inference Engine

US Patent:
2019024, Aug 8, 2019
Filed:
Dec 19, 2018
Appl. No.:
16/226564
Inventors:
- Santa Clara CA, US
Ulf HANEBUTTE - Gig Harbor WA, US
Senad DURAKOVIC - Palo Alto CA, US
Hamid Reza GHASEMI - Sunnyvale CA, US
Chia-Hsin CHEN - Santa Clara CA, US
International Classification:
G06N 5/04
G06N 20/00
G06F 17/16
Abstract:
A programmable hardware system for machine learning (ML) includes a core and an inference engine. The core receives commands from a host. The commands are in a first instruction set architecture (ISA) format. The core divides the commands into a first set for performance-critical operations, in the first ISA format, and a second set of performance non-critical operations, in the first ISA format. The core executes the second set to perform the performance non-critical operations of the ML operations and streams the first set to inference engine. The inference engine generates a stream of the first set of commands in a second ISA format based on the first set of commands in the first ISA format. The first set of commands in the second ISA format programs components within the inference engine to execute the ML operations to infer data.

Architecture For Dense Operations In Machine Learning Inference Engine

US Patent:
2019024, Aug 8, 2019
Filed:
Dec 19, 2018
Appl. No.:
16/226550
Inventors:
- Santa Clara CA, US
Ulf HANEBUTTE - Gig Harbor WA, US
Senad DURAKOVIC - Palo Alto CA, US
Hamid Reza GHASEMI - Sunnyvale CA, US
Chia-Hsin CHEN - Santa Clara CA, US
International Classification:
G06N 20/00
G06N 5/04
G06F 17/16
G06F 12/0862
Abstract:
A processing unit of an inference engine for machine learning (ML) includes a first, a second, and a third register, and a matrix multiplication block. The first register receives a first stream of data associated with a first matrix data that is read only once. The second register receives a second stream of data associated with a second matrix data that is read only once. The matrix multiplication block performs a multiplication operation based on data from the first register and the second register resulting in an output matrix. A row associated with the first matrix is maintained while rows associated with the second matrix is fed to the matrix multiplication block to perform a multiplication operation. The process is repeated for each row of the first matrix. The third register receives the output matrix from the matrix multiplication block and stores the output matrix.

NOTICE: You may not use PeopleBackgroundCheck or the information it provides to make decisions about employment, credit, housing or any other purpose that would require Fair Credit Reporting Act (FCRA) compliance. PeopleBackgroundCheck is not a Consumer Reporting Agency (CRA) as defined by the FCRA and does not provide consumer reports.