The Quantum Exact Simulation Toolkit v4.0.0
Loading...
Searching...
No Matches
matrices.cpp
1/** @file
2 * Definitions of API matrix data structures, and their getters
3 * and setters, as well as reporting utilities.
4 *
5 * This file defines several layers of initialisation of complex
6 * matrices, as explained in the header file.
7 *
8 * @author Tyson Jones
9 */
10
11#include "quest/include/matrices.h"
12#include "quest/include/environment.h"
13#include "quest/include/types.h"
14
15#include "quest/src/core/validation.hpp"
16#include "quest/src/core/autodeployer.hpp"
17#include "quest/src/core/utilities.hpp"
18#include "quest/src/core/localiser.hpp"
19#include "quest/src/core/printer.hpp"
20#include "quest/src/core/bitwise.hpp"
21#include "quest/src/core/fastmath.hpp"
22#include "quest/src/core/memory.hpp"
23#include "quest/src/comm/comm_config.hpp"
24#include "quest/src/comm/comm_routines.hpp"
25#include "quest/src/cpu/cpu_config.hpp"
26#include "quest/src/gpu/gpu_config.hpp"
27#include "quest/src/cpu/cpu_subroutines.hpp"
28
29#include <iostream>
30#include <cstdlib>
31#include <string>
32#include <vector>
33
34using std::vector;
35
36
37
38/*
39 * FIZED-SIZE MATRIX VECTOR GETTERS
40 *
41 * enabling getCompMatr1 (etc) to receive vectors, in addition
42 * to the existing pointer and array overloads in the header.
43 * These are necessarily defined here in the source, unlike the
44 * other header functions, so that they may use private valdiation.
45 */
46
47
48CompMatr1 getCompMatr1(vector<vector<qcomp>> in) {
49 validate_matrixNumNewElems(1, in, __func__);
50
51 qcomp* rowPtrs[] = {in[0].data(), in[1].data()};
52 return getCompMatr1(rowPtrs);
53}
54
55CompMatr2 getCompMatr2(vector<vector<qcomp>> in) {
56 validate_matrixNumNewElems(2, in, __func__);
57
58 qcomp* rowPtrs[] = {in[0].data(), in[1].data(), in[2].data(), in[3].data()};
59 return getCompMatr2(rowPtrs);
60}
61
62
63DiagMatr1 getDiagMatr1(vector<qcomp> in) {
64 validate_matrixNumNewElems(1, in, __func__);
65
66 return getDiagMatr1(in.data());
67}
68
69DiagMatr2 getDiagMatr2(vector<qcomp> in) {
70 validate_matrixNumNewElems(2, in, __func__);
71
72 return getDiagMatr2(in.data());
73}
74
75
76
77/*
78 * PRIVATE VARIABLE-SIZE MATRIX UTILITIES
79 */
80
81
82// type T can be CompMatr, DiagMatr or FullStateDiagMatr
83template <class T>
84void freeHeapMatrix(T matr) {
85
86 // WARNING: this is not overwriting any freed pointers with null,
87 // since the caller's struct is not changed. This is fine here,
88 // but would be an issue if the struct contained nested pointers,
89 // since caller would not know an outer pointer was freed and
90 // ergo that it should not be enumerated (to check/free inner ptr)
91
92 // free the 1D or 2D matrix - safe even if nullptr
93 if constexpr (util_isDenseMatrixType<T>()) {
94 cpu_deallocMatrixWrapper(matr.cpuElems);
95 cpu_deallocArray(matr.cpuElemsFlat);
96 } else
97 cpu_deallocArray(matr.cpuElems);
98
99 // we avoid invoking a GPU function in non-GPU mode
100 auto gpuPtr = util_getGpuMemPtr(matr);
101 if (mem_isAllocated(gpuPtr))
102 gpu_deallocArray(gpuPtr);
103
104 // free the teeny tiny heap flags
105 util_deallocEpsilonSensitiveHeapFlag(matr.isApproxUnitary);
106 util_deallocEpsilonSensitiveHeapFlag(matr.isApproxHermitian);
107 cpu_deallocHeapFlag(matr.wasGpuSynced);
108
109 // only diagonal matrices (which can be raised to
110 // exponents) need their negativity/zeroness checked
111 if constexpr (!util_isDenseMatrixType<T>()) {
112 util_deallocEpsilonSensitiveHeapFlag(matr.isApproxNonZero);
113 cpu_deallocHeapFlag(matr.isStrictlyNonNegative);
114 }
115}
116
117
118// type T can be CompMatr, DiagMatr or FullStateDiagMatr
119template <class T>
120bool didAnyLocalAllocsFail(T matr) {
121
122 // god help us if these single-integer malloc failed
123 if (!mem_isAllocated(matr.isApproxUnitary)) return true;
124 if (!mem_isAllocated(matr.isApproxHermitian)) return true;
125 if (!mem_isAllocated(matr.wasGpuSynced)) return true;
126
127 // only diagonal matrices (which can be raised to
128 // exponents) have these addtional fields
129 if constexpr (!util_isDenseMatrixType<T>()) {
130 if (!mem_isAllocated(matr.isApproxNonZero)) return true;
131 if (!mem_isAllocated(matr.isStrictlyNonNegative)) return true;
132 }
133
134 // outer CPU memory should always be allocated
135 if constexpr (util_isDenseMatrixType<T>()) {
136 if (!mem_isAllocated(matr.cpuElemsFlat)) return true;
137 if (!mem_isOuterAllocated(matr.cpuElems)) return true;
138 } else
139 if (!mem_isAllocated(matr.cpuElems)) return true;
140
141 // if GPU memory is not allocated in a GPU environment...
142 bool isGpuAlloc = mem_isAllocated(util_getGpuMemPtr(matr));
143 if (getQuESTEnv().isGpuAccelerated && !isGpuAlloc) {
144
145 // then FullStateDiagMatr GPU alloc failed only if it tried...
146 if constexpr (util_isFullStateDiagMatr<T>()) {
147 if (matr.isGpuAccelerated)
148 return true;
149
150 // but all other matrices always try to alloc, so must have failed
151 } else
152 return true;
153 }
154
155 // otherwise, all pointers were non-NULL and ergo all allocs were successful
156 return false;
157}
158
159
160// type T can be CompMatr, DiagMatr or FullStateDiagMatr
161template <class T>
162void freeAllMemoryIfAnyAllocsFailed(T matr) {
163
164 // ascertain whether any allocs failed on any node
165 bool anyFail = didAnyLocalAllocsFail(matr);
166 if (comm_isInit())
167 anyFail = comm_isTrueOnAllNodes(anyFail);
168
169 // if so, free all heap fields
170 if (anyFail)
171 freeHeapMatrix(matr);
172}
173
174
175// type T can be CompMatr, DiagMatr or FullStateDiagMatr
176template <class T>
177void validateMatrixAllocs(T matr, const char* caller) {
178
179 // free memory before throwing validation error to avoid memory leaks
180 freeAllMemoryIfAnyAllocsFailed(matr);
181 validate_newMatrixAllocs(matr, caller);
182}
183
184
185// type T can be CompMatr, DiagMatr or FullStateDiagMatr
186template <class T>
187void setInitialHeapFlags(T matr) {
188
189 // set initial propreties of the newly created matrix to unknown
190 util_setFlagToUnknown(matr.isApproxUnitary);
191 util_setFlagToUnknown(matr.isApproxHermitian);
192
193 // only diagonal matrices (which can be exponentiated)
194 // have these additional fields
195 if constexpr (!util_isDenseMatrixType<T>()) {
196 util_setFlagToUnknown(matr.isApproxNonZero);
197 util_setFlagToUnknown(matr.isStrictlyNonNegative);
198 }
199
200 // indicate that GPU memory has not yet been synchronised
201 *(matr.wasGpuSynced) = 0;
202}
203
204
205
206/*
207 * VARIABLE-SIZE MATRIX CONSTRUCTORS
208 */
209
210
211extern "C" CompMatr createCompMatr(int numQubits) {
212 validate_envIsInit(__func__);
213 validate_newCompMatrParams(numQubits, __func__);
214
215 // validation ensures these never overflow
216 qindex numRows = powerOf2(numQubits);
217 qindex numElems = numRows * numRows;
218
219 qcomp* cpuMem = cpu_allocArray(numElems); // nullptr if failed
220 qcomp* gpuMem = nullptr;
221 if (getQuESTEnv().isGpuAccelerated)
222 gpuMem = gpu_allocArray(numElems); // nullptr if failed
223
224 // initialise all CompMatr fields inline because most are const
225 CompMatr out = {
226 .numQubits = numQubits,
227 .numRows = numRows,
228
229 // allocate flags in the heap so that struct copies are mutable
230 .isApproxUnitary = util_allocEpsilonSensitiveHeapFlag(), // nullptr if failed
231 .isApproxHermitian = util_allocEpsilonSensitiveHeapFlag(),
232
233 .wasGpuSynced = cpu_allocHeapFlag(), // nullptr if failed
234
235 .cpuElems = cpu_allocAndInitMatrixWrapper(cpuMem, numRows), // nullptr if failed
236 .cpuElemsFlat = cpuMem,
237 .gpuElemsFlat = gpuMem
238 };
239
240 validateMatrixAllocs(out, __func__);
241 setInitialHeapFlags(out);
242 return out;
243}
244
245
246extern "C" DiagMatr createDiagMatr(int numQubits) {
247 validate_envIsInit(__func__);
248 validate_newDiagMatrParams(numQubits, __func__);
249
250 // validation ensures this never overflows
251 qindex numElems = powerOf2(numQubits);
252
253 // initialise all CompMatr fields inline because most are const
254 DiagMatr out = {
255 .numQubits = numQubits,
256 .numElems = numElems,
257
258 // allocate flags in the heap so that struct copies are mutable
259 .isApproxUnitary = util_allocEpsilonSensitiveHeapFlag(), // nullptr if failed
260 .isApproxHermitian = util_allocEpsilonSensitiveHeapFlag(),
261 .isApproxNonZero = util_allocEpsilonSensitiveHeapFlag(),
262 .isStrictlyNonNegative = cpu_allocHeapFlag(), // nullptr if failed
263 .wasGpuSynced = cpu_allocHeapFlag(),
264
265 // 1D CPU memory
266 .cpuElems = cpu_allocArray(numElems), // nullptr if failed
267
268 // 1D GPU memory
269 .gpuElems = (getQuESTEnv().isGpuAccelerated)? gpu_allocArray(numElems) : nullptr // nullptr if failed or not needed
270 };
271
272 validateMatrixAllocs(out, __func__);
273 setInitialHeapFlags(out);
274 return out;
275}
276
277
278FullStateDiagMatr validateAndCreateCustomFullStateDiagMatr(int numQubits, int useDistrib, int useGpuAccel, int useMultithread, const char* caller) {
279 validate_envIsInit(caller);
280 QuESTEnv env = getQuESTEnv();
281
282 // validate parameters before passing them to autodeployer
283 validate_newFullStateDiagMatrParams(numQubits, useDistrib, useGpuAccel, useMultithread, caller);
284
285 // overwrite useDistrib and useGpuAccel if they were left as AUTO_FLAG
286 autodep_chooseFullStateDiagMatrDeployment(numQubits, useDistrib, useGpuAccel, useMultithread, env);
287
288 // validation ensures this never overflows
289 qindex numElems = powerOf2(numQubits);
290 qindex numElemsPerNode = numElems / (useDistrib? env.numNodes : 1); // divides evenly
291
292 FullStateDiagMatr out = {
293
294 .numQubits = numQubits,
295 .numElems = numElems,
296
297 // data deployment configuration; disable distrib if deployed to 1 node
298 .isGpuAccelerated = useGpuAccel,
299 .isMultithreaded = useMultithread,
300 .isDistributed = useDistrib && (env.numNodes > 1),
301 .numElemsPerNode = numElemsPerNode,
302
303 // allocate flags in the heap so that struct copies are mutable
304 .isApproxUnitary = util_allocEpsilonSensitiveHeapFlag(), // nullptr if failed
305 .isApproxHermitian = util_allocEpsilonSensitiveHeapFlag(),
306 .isApproxNonZero = util_allocEpsilonSensitiveHeapFlag(),
307 .isStrictlyNonNegative = cpu_allocHeapFlag(), // nullptr if failed
308 .wasGpuSynced = cpu_allocHeapFlag(),
309
310 // 1D CPU memory
311 .cpuElems = cpu_allocArray(numElemsPerNode), // nullptr if failed
312
313 // 1D GPU memory
314 .gpuElems = (useGpuAccel)? gpu_allocArray(numElemsPerNode) : nullptr, // nullptr if failed or not needed
315 };
316
317 validateMatrixAllocs(out, __func__);
318 setInitialHeapFlags(out);
319 return out;
320}
321
322extern "C" FullStateDiagMatr createCustomFullStateDiagMatr(int numQubits, int useDistrib, int useGpuAccel, int useMultithread) {
323
324 return validateAndCreateCustomFullStateDiagMatr(numQubits, useDistrib, useGpuAccel, useMultithread, __func__);
325}
326
328
329 return validateAndCreateCustomFullStateDiagMatr(numQubits, modeflag::USE_AUTO, modeflag::USE_AUTO, modeflag::USE_AUTO, __func__);
330}
331
332
333
334/*
335 * VARIABLE-SIZE MATRIX SYNC
336 */
337
338
339// type T can be CompMatr, DiagMatr or FullStateDiagMatr
340template <class T>
341void markMatrixAsSynced(T matr) {
342
343 // indicate that the matrix is now permanently GPU synchronised, even
344 // if we are not in GPU-accelerated mode (in which case it's never consulted)
345 *(matr.wasGpuSynced) = 1;
346
347 // indicate that we do not know the revised matrix properties;
348 // we defer establishing that until validation needs to check them
349 util_setFlagToUnknown(matr.isApproxUnitary);
350 util_setFlagToUnknown(matr.isApproxHermitian);
351
352 // only diagonal matrices (which can be exponentiated)
353 // have these additional fields
354 if constexpr (!util_isDenseMatrixType<T>()) {
355 util_setFlagToUnknown(matr.isApproxNonZero);
356 util_setFlagToUnknown(matr.isStrictlyNonNegative);
357 }
358}
359
360
361// type T can be CompMatr, DiagMatr or FullStateDiagMatr
362template <class T>
363void validateAndSyncMatrix(T matr, const char* caller) {
364 validate_matrixFields(matr, caller);
365
366 // optionally overwrite GPU elements with user-modified CPU elements
367 if (mem_isAllocated(util_getGpuMemPtr(matr)))
368 gpu_copyCpuToGpu(matr);
369
370 markMatrixAsSynced(matr);
371}
372
373
374// de-mangled for both C++ and C compatibility
375extern "C" {
376
377 void syncCompMatr(CompMatr matr) { validateAndSyncMatrix(matr, __func__); }
378 void syncDiagMatr(DiagMatr matr) { validateAndSyncMatrix(matr, __func__); }
379 void syncFullStateDiagMatr(FullStateDiagMatr matr) { validateAndSyncMatrix(matr, __func__); }
380
381}
382
383
384
385/*
386 * VARIABLE-SIZE MATRIX DESTRUCTION
387 */
388
389
390// type T can be CompMatr, DiagMatr or FullStateDiagMatr
391template <class T>
392void validateAndDestroyMatrix(T matrix, const char* caller) {
393 validate_matrixFields(matrix, caller);
394
395 freeHeapMatrix(matrix);
396}
397
398
399// de-mangled for C++ and C compatibility
400extern "C" {
401
402 void destroyCompMatr(CompMatr matr) { validateAndDestroyMatrix(matr, __func__); }
403 void destroyDiagMatr(DiagMatr matr) { validateAndDestroyMatrix(matr, __func__); }
404 void destroyFullStateDiagMatr(FullStateDiagMatr matr) { validateAndDestroyMatrix(matr, __func__); }
405
406}
407
408
409
410/*
411 * VARIABLE-SIZE MATRIX SETTERS VIA POINTERS
412 */
413
414
415// type T can be qcomp** or vector<vector<qcomp>>, but qcomp(*)[] is handled by header
416template <typename T>
417void setAndSyncDenseMatrElems(CompMatr out, T elems) {
418
419 // copy elems into matrix's CPU memory
420 cpu_copyMatrix(out.cpuElems, elems, out.numRows);
421
422 // overwrite GPU elements; validation gauranteed to pass
423 syncCompMatr(out);
424}
425
426
427extern "C" void setCompMatr(CompMatr out, qcomp** in) {
428 validate_matrixFields(out, __func__);
429 validate_matrixNewElemsPtrNotNull(in, out.numRows, __func__);
430
431 setAndSyncDenseMatrElems(out, in);
432}
433
434
435extern "C" void setDiagMatr(DiagMatr out, qcomp* in) {
436 validate_matrixFields(out, __func__);
437 validate_matrixNewElemsPtrNotNull(in, __func__);
438
439 // overwrite CPU memory
440 cpu_copyArray(out.cpuElems, in, out.numElems);
441
442 // overwrite GPU elements; validation gauranteed to pass
443 syncDiagMatr(out);
444}
445
446
447extern "C" void setFullStateDiagMatr(FullStateDiagMatr out, qindex startInd, qcomp* in, qindex numElems) {
448 validate_matrixFields(out, __func__);
449 validate_fullStateDiagMatrNewElems(out, startInd, numElems, __func__);
450 validate_matrixNewElemsPtrNotNull(in, __func__);
451
452 // overwrites both the CPU and GPU memory (if it exists), maintaining consistency.
453 // note that cpu_copyArray() isn't called directly here like setDiagMatr() above
454 // because we must handle when 'out' is and isn't distributed
455 localiser_fullstatediagmatr_setElems(out, startInd, in, numElems);
456
457 // even though we have not necessarily overwritten every element, we must mark
458 // the matrix as synced so that it can be subsequently used without error
459 markMatrixAsSynced(out);
460}
461
462
463
464/*
465 * VARIABLE-SIZE MATRIX SETTERS VIA VECTORS
466 */
467
468
469void setCompMatr(CompMatr out, vector<vector<qcomp>> in) {
470 validate_matrixFields(out, __func__);
471 validate_matrixNumNewElems(out.numQubits, in, __func__);
472
473 setAndSyncDenseMatrElems(out, in);
474}
475
476
477void setDiagMatr(DiagMatr out, vector<qcomp> in) {
478 validate_matrixFields(out, __func__);
479 validate_matrixNumNewElems(out.numQubits, in, __func__); // validates 'in' dim
480
481 setDiagMatr(out, in.data()); // harmessly re-validates
482}
483
484
485void setFullStateDiagMatr(FullStateDiagMatr out, qindex startInd, vector<qcomp> in) {
486
487 setFullStateDiagMatr(out, startInd, in.data(), in.size());
488}
489
490
491// no bespoke array functions are necessary for diagonal matrices initialisation,
492// since passed arrays automatically decay to pointers
493
494
495
496/*
497 * VARIABLE-SIZE MATRIX SETTERS VIA LITERALS
498 *
499 * Only the C++ versions are defined here, while the C versions are macros
500 * defined in the header. Note the C++ versions themselves are entirely
501 * superfluous and merely call the above vector setters, but we still define
502 * them for API consistency, and we additionally validate the superfluous
503 * additional parameters they pass.
504 */
505
506
507void setInlineCompMatr(CompMatr matr, int numQb, vector<vector<qcomp>> in) {
508 validate_matrixFields(matr, __func__);
509 validate_matrixNumQubitsMatchesParam(matr.numQubits, numQb, __func__);
510 validate_matrixNumNewElems(matr.numQubits, in, __func__);
511
512 setAndSyncDenseMatrElems(matr, in);
513}
514
515void setInlineDiagMatr(DiagMatr matr, int numQb, vector<qcomp> in) {
516 validate_matrixFields(matr, __func__);
517 validate_matrixNumQubitsMatchesParam(matr.numQubits, numQb, __func__);
518 validate_matrixNumNewElems(matr.numQubits, in, __func__);
519
520 setDiagMatr(matr, in.data()); // validation gauranteed to pass
521}
522
523void setInlineFullStateDiagMatr(FullStateDiagMatr matr, qindex startInd, qindex numElems, vector<qcomp> in) {
524 validate_matrixFields(matr, __func__);
525 validate_declaredNumElemsMatchesVectorLength(numElems, in.size(), __func__);
526 validate_fullStateDiagMatrNewElems(matr, startInd, numElems, __func__);
527
528 setFullStateDiagMatr(matr, startInd, in); // validation gauranteed to pass
529}
530
531
532
533/*
534 * VARIABLE-SIZE MATRIX INLINE-SETTER CONSTRUCTORS
535 *
536 * Only the C++ versions are defined here; the C versions are header macros
537 */
538
539
540CompMatr createInlineCompMatr(int numQb, vector<vector<qcomp>> elems) {
541 validate_envIsInit(__func__);
542 validate_newCompMatrParams(numQb, __func__);
543 validate_matrixNumNewElems(numQb, elems, __func__);
544
545 // pre-validation gauranteed to pass, but malloc failures will trigger an error
546 // message specific to 'createCompMatr', rather than this 'inline' version. Alas!
547 CompMatr matr = createCompMatr(numQb);
548 setAndSyncDenseMatrElems(matr, elems);
549 return matr;
550}
551
552
553DiagMatr createInlineDiagMatr(int numQb, vector<qcomp> elems) {
554 validate_envIsInit(__func__);
555 validate_newDiagMatrParams(numQb, __func__);
556 validate_matrixNumNewElems(numQb, elems, __func__);
557
558 // pre-validation gauranteed to pass, but malloc failures will trigger an error
559 // message specific to 'createCompMatr', rather than this 'inline' version. Alas!
560 DiagMatr matr = createDiagMatr(numQb);
561 setDiagMatr(matr, elems.data()); // validation gauranteed to pass
562 return matr;
563}
564
565
566
567/*
568 * EXPOSING SOME SETTER VALIDATION TO HEADER
569 *
570 * Some setters are necessarily defined in the header, because they accept
571 * C-only VLAs which need to be cast into pointers before being passed to
572 * this C++ backend (which does not support VLA). These setters need their
573 * validators exposed, though we cannot expose the entirety of validation.hpp
574 * because it cannot be parsed by C; so we here wrap specific functions.
575 */
576
577
578extern "C" {
579
580 void _validateNewNestedElemsPtrNotNull(qcomp** ptrs, int numQubits, const char* caller) {
581
582 validate_matrixNewElemsPtrNotNull(ptrs, powerOf2(numQubits), caller);
583 }
584 void _validateNewElemsPtrNotNull(qcomp* ptr, const char* caller) {
585
586 validate_matrixNewElemsPtrNotNull(ptr, caller);
587 }
588
589 void _validateParamsToSetCompMatrFromArr(CompMatr matr) {
590
591 validate_matrixFields(matr, "setCompMatr");
592 }
593
594 void _validateParamsToSetInlineCompMatr(CompMatr matr, int numQb) {
595
596 const char* caller = "setInlineCompMatr";
597 validate_matrixFields(matr, caller);
598 validate_matrixNumQubitsMatchesParam(matr.numQubits, numQb, caller);
599 }
600
601 void _validateParamsToSetInlineDiagMatr(DiagMatr matr, int numQb) {
602
603 const char* caller = "setInlineDiagMatr";
604 validate_matrixFields(matr, caller);
605 validate_matrixNumQubitsMatchesParam(matr.numQubits, numQb, caller);
606 }
607
608 void _validateParamsToSetInlineFullStateDiagMatr(FullStateDiagMatr matr, qindex startInd, qindex numElems) {
609
610 const char* caller = "setInlineFullStateDiagMatr";
611 validate_matrixFields(matr, caller);
612 validate_fullStateDiagMatrNewElems(matr, startInd, numElems, caller);
613 }
614
615 void _validateParamsToCreateInlineCompMatr(int numQb) {
616
617 const char* caller = "createInlineCompMatr";
618 validate_envIsInit(caller);
619 validate_newCompMatrParams(numQb, caller);
620 }
621
622 void _validateParamsToCreateInlineDiagMatr(int numQb) {
623
624 const char* caller = "createInlineDiagMatr";
625 validate_envIsInit(caller);
626 validate_newDiagMatrParams(numQb, caller);
627 }
628
629}
630
631
632
633/*
634 * SPECIAL CREATORS AND SETTERS
635 */
636
637
638extern int paulis_getIndOfLefmostNonIdentityPauli(PauliStrSum sum);
639
640
642 validate_matrixFields(out, __func__);
643 validate_pauliStrSumFields(in, __func__);
644 validate_pauliStrSumCanInitMatrix(out, in, __func__);
645
646 // permit 'in' to be non-Hermitian since it does not determine 'out' unitarity
647
648 // unlike other FullStateDiagMatr initialisers, we employ an accelerated
649 // backend since the input data 'in' is expectedly significantly smaller
650 // than the created data in 'out', making parallelisation worthwhile as
651 // the memory-movement costs of copying 'in' to a GPU are small
652 localiser_fullstatediagmatr_setElemsToPauliStrSum(out, in);
653
654 markMatrixAsSynced(out);
655}
656
657
659 validate_pauliStrSumFields(in, __func__);
660
661 // ensure createFullStateDiagMatr() below succeeds (so if not, that thrower name is correct)
662 int numQubits = 1 + paulis_getIndOfLefmostNonIdentityPauli(in);
663 validate_newFullStateDiagMatrParams(numQubits, modeflag::USE_AUTO, modeflag::USE_AUTO, modeflag::USE_AUTO, __func__);
664
665 // permit 'in' to be non-Hermitian since it does not determine 'out' unitarity
666
668 localiser_fullstatediagmatr_setElemsToPauliStrSum(out, in);
669 markMatrixAsSynced(out);
670 return out;
671}
672
673
674extern "C" void setDiagMatrFromMultiVarFunc(DiagMatr out, qcomp (*callbackFunc)(qindex*), int* numQubitsPerVar, int numVars, int areSigned) {
675 validate_matrixFields(out, __func__);
676 validate_multiVarFuncQubits(out.numQubits, numQubitsPerVar, numVars, __func__);
677 validate_funcVarSignedFlag(areSigned, __func__);
678
679 vector<qindex> varValues(numVars);
680
681 // set each element of the diagonal in-turn; user's callback might not be thread-safe
682 for (qindex elemInd=0; elemInd<out.numElems; elemInd++) {
683 fast_getSubQuregValues(elemInd, numQubitsPerVar, numVars, areSigned, varValues.data());
684
685 // call user function, and update only the CPU elems
686 out.cpuElems[elemInd] = callbackFunc(varValues.data());
687 }
688
689 // overwrite all GPU elems
690 syncDiagMatr(out);
691}
692
693
694extern "C" void setFullStateDiagMatrFromMultiVarFunc(FullStateDiagMatr out, qcomp (*callbackFunc)(qindex*), int* numQubitsPerVar, int numVars, int areSigned) {
695 validate_matrixFields(out, __func__);
696 validate_multiVarFuncQubits(out.numQubits, numQubitsPerVar, numVars, __func__);
697 validate_funcVarSignedFlag(areSigned, __func__);
698
699 // we assume callbackFunc is thread-safe (!!!!) and possibly use multithreading, but never
700 // GPU acceleration, since we cannot invoke user callback functions from GPU kernels
701 cpu_fullstatediagmatr_setElemsFromMultiVarFunc(out, callbackFunc, numQubitsPerVar, numVars, areSigned);
702
703 // overwrite all GPU elems
705}
706
707
708extern "C" void setDiagMatrFromMultiDimLists(DiagMatr out, void* lists, int* numQubitsPerDim, int numDims) {
709 validate_matrixFields(out, __func__);
710 validate_multiVarFuncQubits(out.numQubits, numQubitsPerDim, numDims, __func__);
711
712 vector<qindex> listInds(numDims);
713
714 // set each element of the diagonal in-turn, which is embarrassingly parallel,
715 // although we do not parallelise - the DiagMatr is intendedly small
716 for (qindex elemInd=0; elemInd<out.numElems; elemInd++) {
717
718 // nested list indices = unsigned integer values of variables
719 fast_getSubQuregValues(elemInd, numQubitsPerDim, numDims, false, listInds.data());
720
721 // update only the CPU elems
722 out.cpuElems[elemInd] = util_getElemFromNestedPtrs(lists, listInds.data(), numDims);
723 }
724
725 // overwrite all GPU elems
726 syncDiagMatr(out);
727}
728
729
730extern "C" void setFullStateDiagMatrFromMultiDimLists(FullStateDiagMatr out, void* lists, int* numQubitsPerDim, int numDims) {
731 validate_matrixFields(out, __func__);
732 validate_multiVarFuncQubits(out.numQubits, numQubitsPerDim, numDims, __func__);
733
734 // possibly use multithreading, but never GPU acceleration, due to the
735 // arbitrarily nested nature of the input lists
736 cpu_fullstatediagmatr_setElemsFromMultiDimLists(out, lists, numQubitsPerDim, numDims);
737
738 // overwrite all GPU elems
740}
741
742
743
744/*
745 * MATRIX REPORTERS
746 */
747
748
749// type T can be CompMatr1, CompMatr2, CompMatr, DiagMatr1, DiagMatr2, DiagMatr, FullStateDiagMatr
750template<class T>
751void validateAndPrintMatrix(T matr, const char* caller) {
752 validate_matrixFields(matr, caller);
753 validate_numReportedNewlinesAboveZero(__func__); // because trailing newline mandatory
754
755 // syncable matrices must be synced before reporting (though only CPU elems are printed)
756 if constexpr (util_isHeapMatrixType<T>())
757 validate_matrixIsSynced(matr, caller);
758
759 // calculate the total memory (in bytes) consumed by the matrix on each
760 // node, which will depend on whether the matrix is distributed, and
761 // includes the size of the matrix struct itself. Note that GPU memory
762 // is not included since it is less than or equal to the CPU memory, and
763 // occupies different memory spaces, confusing capacity accounting
764 int numNodes = (util_isDistributedMatrix(matr))? comm_getNumNodes() : 1;
765 size_t elemMem = mem_getLocalMatrixMemoryRequired(matr.numQubits, util_isDenseMatrixType<T>(), numNodes);
766 size_t structMem = sizeof(matr);
767
768 // struct memory includes fixed-size arrays (qcomp[][]), so we undo double-counting
769 if (util_isFixedSizeMatrixType<T>())
770 structMem -= elemMem;
771
772 size_t numBytesPerNode = elemMem + structMem;
773 print_header(matr, numBytesPerNode);
774 print_elems(matr);
775
776 // exclude mandatory newline above
777 print_oneFewerNewlines();
778}
779
780
781// all reporters are C and C++ accessible, so are de-mangled
782extern "C" {
783
784 void reportCompMatr1(CompMatr1 matr) { validateAndPrintMatrix(matr, __func__); }
785 void reportCompMatr2(CompMatr2 matr) { validateAndPrintMatrix(matr, __func__); }
786 void reportCompMatr (CompMatr matr) { validateAndPrintMatrix(matr, __func__); }
787 void reportDiagMatr1(DiagMatr1 matr) { validateAndPrintMatrix(matr, __func__); }
788 void reportDiagMatr2(DiagMatr2 matr) { validateAndPrintMatrix(matr, __func__); }
789 void reportDiagMatr (DiagMatr matr) { validateAndPrintMatrix(matr, __func__); }
790 void reportFullStateDiagMatr(FullStateDiagMatr matr) { validateAndPrintMatrix(matr, __func__); }
791
792}
QuESTEnv getQuESTEnv()
FullStateDiagMatr createCustomFullStateDiagMatr(int numQubits, int useDistrib, int useGpuAccel, int useMultithread)
Definition matrices.cpp:322
DiagMatr createInlineDiagMatr(int numQb, std::vector< qcomp > elems)
FullStateDiagMatr createFullStateDiagMatr(int numQubits)
Definition matrices.cpp:327
CompMatr createInlineCompMatr(int numQb, std::vector< std::vector< qcomp > > elems)
CompMatr createCompMatr(int numQubits)
Definition matrices.cpp:211
DiagMatr createDiagMatr(int numQubits)
Definition matrices.cpp:246
FullStateDiagMatr createFullStateDiagMatrFromPauliStrSum(PauliStrSum in)
Definition matrices.cpp:658
void destroyDiagMatr(DiagMatr matr)
Definition matrices.cpp:403
void destroyFullStateDiagMatr(FullStateDiagMatr matr)
Definition matrices.cpp:404
void destroyCompMatr(CompMatr matr)
Definition matrices.cpp:402
DiagMatr1 getDiagMatr1(std::vector< qcomp > in)
CompMatr1 getCompMatr1(std::vector< std::vector< qcomp > > in)
CompMatr2 getCompMatr2(std::vector< std::vector< qcomp > > in)
DiagMatr2 getDiagMatr2(std::vector< qcomp > in)
void reportCompMatr2(CompMatr2 matr)
Definition matrices.cpp:785
void reportDiagMatr1(DiagMatr1 matr)
Definition matrices.cpp:787
void reportCompMatr(CompMatr matr)
Definition matrices.cpp:786
void reportDiagMatr2(DiagMatr2 matr)
Definition matrices.cpp:788
void reportCompMatr1(CompMatr1 matr)
Definition matrices.cpp:784
void reportDiagMatr(DiagMatr matr)
Definition matrices.cpp:789
void reportFullStateDiagMatr(FullStateDiagMatr matr)
Definition matrices.cpp:790
void setInlineDiagMatr(DiagMatr matr, int numQb, std::vector< qcomp > in)
void setFullStateDiagMatrFromMultiDimLists(FullStateDiagMatr out, void *lists, int *numQubitsPerDim, int numDims)
Definition matrices.cpp:730
void setDiagMatrFromMultiDimLists(DiagMatr out, void *lists, int *numQubitsPerDim, int numDims)
Definition matrices.cpp:708
void setFullStateDiagMatrFromMultiVarFunc(FullStateDiagMatr out, qcomp(*callbackFunc)(qindex *), int *numQubitsPerVar, int numVars, int areSigned)
Definition matrices.cpp:694
void setInlineCompMatr(CompMatr matr, int numQb, std::vector< std::vector< qcomp > > in)
void setDiagMatr(DiagMatr out, qcomp *in)
Definition matrices.cpp:435
void setCompMatr(CompMatr out, qcomp **in)
Definition matrices.cpp:427
void setFullStateDiagMatr(FullStateDiagMatr out, qindex startInd, qcomp *in, qindex numElems)
Definition matrices.cpp:447
void setDiagMatrFromMultiVarFunc(DiagMatr out, qcomp(*callbackFunc)(qindex *), int *numQubitsPerVar, int numVars, int areSigned)
Definition matrices.cpp:674
void setFullStateDiagMatrFromPauliStrSum(FullStateDiagMatr out, PauliStrSum in)
Definition matrices.cpp:641
void setInlineFullStateDiagMatr(FullStateDiagMatr matr, qindex startInd, qindex numElems, std::vector< qcomp > in)
void syncFullStateDiagMatr(FullStateDiagMatr matr)
Definition matrices.cpp:379
void syncDiagMatr(DiagMatr matr)
Definition matrices.cpp:378
void syncCompMatr(CompMatr matr)
Definition matrices.cpp:377