max_stars_count
int64
301
224k
text
stringlengths
6
1.05M
token_count
int64
3
727k
634
<filename>modules/desktop-swt/desktop-swt-ui-impl/src/main/java/org/eclipse/nebula/cwt/svg/SvgGraphic.java /**************************************************************************** * Copyright (c) 2008, 2009 <NAME> * * This program and the accompanying materials * are made available under the terms of the Eclipse Public License 2.0 * which accompanies this distribution, and is available at * https://www.eclipse.org/legal/epl-2.0/ * * SPDX-License-Identifier: EPL-2.0 * * Contributors: * <NAME> <<EMAIL>> - initial API and implementation *****************************************************************************/ package org.eclipse.nebula.cwt.svg; import org.eclipse.nebula.cwt.svg.SvgPaint.PaintType; import org.eclipse.swt.SWT; import org.eclipse.swt.graphics.GC; import org.eclipse.swt.graphics.Transform; /** * A base abstract class for all types of svg elements which can be * applied (painted) to a graphics context. These may be shapes which * can be painted directly, or containers which will paint their children. */ public abstract class SvgGraphic extends SvgElement { String title; String description; SvgFill fill; SvgStroke stroke; SvgTransform transform; SvgGraphic(SvgContainer container, String id) { super(container, id); fill = new SvgFill(this); stroke = new SvgStroke(this); } /** * Apply this svg graphic to the given graphics context. * <p>Note that to support the rather abstract structure of svg, * each time this method is called all transformations and css properties * to be calculated and applied. If this is a shape, it will be * painted to the graphics context. Containers will recursively * make this call on their children.</p> * @param gc the gc to use in all graphics operations */ public abstract void apply(GC gc); /** * Returns the value of the <code>desc</code> element that is a child of this svg element. * If there is no <code>desc</code> element that is a direct decendent of this element, null * is returned. * @return the <code>desc</code> of this svg element */ public String getDescription() { return (description == null) ? null : description; } SvgFill getFill() { SvgFill df = new SvgFill(this); for(SvgElement el : getAncestry(this)) { if(el instanceof SvgGraphic) { SvgFill tmp = ((SvgGraphic) el).fill; if(tmp.type != null) { df.type = tmp.type; } if(tmp.color != null) { df.color = tmp.color; } if(tmp.linkId != null) { df.linkId = tmp.linkId; } if(tmp.opacity != null) { df.opacity = tmp.opacity; } if(tmp.rule != null) { df.rule = tmp.rule; } } } if(df.type == null) { df.type = PaintType.Color; } if(df.color == null) { df.color = 0; } if(df.opacity == null) { df.opacity = 1f; } if(df.rule == null) { df.rule = SWT.FILL_EVEN_ODD; } return df; } SvgStroke getStroke() { SvgStroke ds = new SvgStroke(this); for(SvgElement el : getAncestry(this)) { if(el instanceof SvgGraphic) { SvgStroke tmp = ((SvgGraphic) el).stroke; if(tmp.type != null) { ds.type = tmp.type; } if(tmp.color != null) { ds.color = tmp.color; } if(tmp.linkId != null) { ds.linkId = tmp.linkId; } if(tmp.opacity != null) { ds.opacity = tmp.opacity; } if(tmp.width != null) { ds.width = tmp.width; } if(tmp.lineCap != null) { ds.lineCap = tmp.lineCap; } if(tmp.lineJoin != null) { ds.lineJoin = tmp.lineJoin; } } } if(ds.type == null) { ds.type = PaintType.None; } if(ds.type != PaintType.None) { if(ds.color == null) { ds.color = 0; } if(ds.opacity == null) { ds.opacity = 1f; } if(ds.width == null) { ds.width = 1f; } if(ds.lineCap == null) { ds.lineCap = SWT.CAP_FLAT; } if(ds.lineJoin == null) { ds.lineJoin = SWT.JOIN_MITER; } } return ds; } Transform getTransform(GC gc) { Transform t = new Transform(gc.getDevice()); gc.getTransform(t); for(SvgElement el : getAncestry(this)) { if(el instanceof SvgFragment) { SvgTransform st = ((SvgFragment) el).boundsTransform; if(!st.isIdentity()) { Transform tmp = new Transform(gc.getDevice()); tmp.setElements(st.data[0], st.data[1], st.data[2], st.data[3], st.data[4], st.data[5]); t.multiply(tmp); tmp.dispose(); } } else if(el instanceof SvgGraphic) { SvgTransform st = ((SvgGraphic) el).transform; while(st != null) { if(!st.isIdentity()) { Transform tmp = new Transform(gc.getDevice()); tmp.setElements(st.data[0], st.data[1], st.data[2], st.data[3], st.data[4], st.data[5]); t.multiply(tmp); tmp.dispose(); } st = st.next; } } } return t; } /** * Returns the value of the <code>title</code> element that is a child of this svg element. * If there is no <code>title</code> element that is a direct decendent of this element, null * is returned. * @return the <code>title</code> of this svg element */ public String getTitle() { return (title == null) ? null : title; } }
2,131
587
<gh_stars>100-1000 /* * Copyright (C) 2013 salesforce.com, inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.auraframework.impl.source.file; import java.io.File; import java.io.FileInputStream; import java.io.FileNotFoundException; import java.io.IOException; import java.io.InputStreamReader; import java.io.Reader; import java.io.StringReader; import java.io.StringWriter; import java.util.Map; import org.auraframework.def.DefDescriptor; import org.auraframework.def.Definition; import org.auraframework.impl.source.AbstractSourceImpl; import org.auraframework.system.Parser.Format; import org.auraframework.system.Source; import org.auraframework.system.TextSource; import org.auraframework.throwable.AuraRuntimeException; import org.auraframework.util.IOUtil; import org.auraframework.util.text.Hash; import org.auraframework.util.text.HashingReader; import com.google.common.collect.ImmutableMap; public class FileSource<D extends Definition> extends AbstractSourceImpl<D> implements TextSource<D> { private final File file; private final long lastModified; private final Hash hash; private volatile String contents; /** * The 'normal' constructor for a file source. * * This constructor gets both the file path and the mime type from the file given. * The file must exist prior to constructing the source. Behaviour is undefined * if the file does not exist, please do not test this case. * * @param descriptor the descriptor for the source. * @param file the file that contains the source. */ public FileSource(DefDescriptor<D> descriptor, File file) throws IOException { super(descriptor, getFilePath(file), getMimeTypeFromExtension(file.getName())); this.file = file; this.lastModified = file.lastModified(); this.hash = Hash.createPromise(); } /** * A 'format' based constructor. * * This constructor gets both the file path and the mime type from the file given. * The file must exist prior to constructing the source. Behaviour is undefined * if the file does not exist, please do not test this case. * * @param descriptor the descriptor for the source. * @param file the file that contains the source. */ public FileSource(DefDescriptor<D> descriptor, File file, Format format) { super(descriptor, getFilePath(file), format); this.file = file; this.lastModified = file.lastModified(); this.hash = Hash.createPromise(); } @Override public Reader getReader() { return new StringReader(getContents()); } @Override public long getLastModified() { return lastModified; } @Override public String getHash() { if (!hash.isSet()) { throw new RuntimeException("No hash set"); } return hash.toString(); } public static String getFilePath(File file) { try { if (!file.exists()) { throw new AuraRuntimeException("File does not exist: " + file.getPath()); } return file.getCanonicalPath(); } catch (Exception e) { throw new AuraRuntimeException(e); } } /** * @see Source#getContents() */ @Override public String getContents() { if (contents != null) { return contents; } synchronized (this) { if (contents != null) { return contents; } try { Reader reader = new InputStreamReader(new FileInputStream(file), "UTF8"); reader = new HashingReader(reader, hash); StringWriter sw = new StringWriter(); IOUtil.copyStream(reader, sw); contents = sw.toString(); } catch (FileNotFoundException e) { throw new AuraRuntimeException(e); } catch (IOException e) { throw new AuraRuntimeException(e); } } return contents; } private final static Map<String,String> mimeMap; static { mimeMap = new ImmutableMap.Builder<String,String>() .put("js", MIME_JAVASCRIPT) .put("html", MIME_HTML) .put("cmp", MIME_XML) .put("lib", MIME_XML) .put("app", MIME_XML) .put("css", MIME_CSS) .put("evt", MIME_XML) .put("intf", MIME_XML) .put("svg", MIME_SVG) .build(); } /** * Get an aura mime type based on the extension. * * This is used to figure out mime types for files. * * @param name the file name to use. * @return a mime type for the file, "X-application/unknown" for anything we don't understand. */ public static String getMimeTypeFromExtension(String name) { int idx = name.lastIndexOf('.'); String mimetype = null; if (idx != -1) { String ext = name.substring(idx+1); mimetype = mimeMap.get(ext); } if (mimetype != null) { return mimetype; } return "X-application/unknown"; } }
2,299
351
#!/usr/bin/env python3 from tryalgo.dijkstra import dijkstra from tryalgo.graph import listlist_and_matrix_to_listdict graph = [[1, 3], [0, 2, 3], [1, 4, 5], [0, 1, 5, 6], [2, 7], [2, 3, 7, 8], [3, 8, 9], [4, 5, 10], [5, 6, 10], [6, 11], [7, 8], [9]] _ = None # 0 1 2 3 4 5 6 7 8 9 10 11 weights = [[_, 1, _, 4, _, _, _, _, _, _, _, _], # 0 [1, _, 1, 3, _, _, _, _, _, _, _, _], # 1 [_, 1, _, _, 3, 8, _, _, _, _, _, _], # 2 [4, 3, _, _, _, 2, 2, _, _, _, _, _], # 3 [_, _, 3, _, _, _, _, 1, _, _, _, _], # 4 [_, _, 8, 2, _, _, _, 2, 7, _, _, _], # 5 [_, _, _, 2, _, _, _, _, 3, 2, _, _], # 6 [_, _, _, _, 1, 2, _, _, _, _, 3, _], # 7 [_, _, _, _, _, 7, 3, _, _, _, 2, _], # 8 [_, _, _, _, _, _, 2, _, _, _, _, 1], # 9 [_, _, _, _, _, _, _, 3, 2, _, _, _], #10 [_, _, _, _, _, _, _, _, _, 1, _, _]] #11 dist, prec = dijkstra(graph, weights, source=0) print(dist[10]) print("%i %i %i %i %i %i" % (10, prec[10], prec[prec[10]], prec[prec[prec[10]]], prec[prec[prec[prec[10]]]], prec[prec[prec[prec[prec[10]]]]])) sparse_graph = listlist_and_matrix_to_listdict(weights) # provides the same behavior dist, prec = dijkstra(graph, weights, source=0) print(dist[10]) print("%i %i %i %i %i %i" % (10, prec[10], prec[prec[10]], prec[prec[prec[10]]], prec[prec[prec[prec[10]]]], prec[prec[prec[prec[prec[10]]]]]))
959
852
#ifndef DataFormats_ParticleFlowReco_PFCluster_h #define DataFormats_ParticleFlowReco_PFCluster_h #include <iostream> #include <vector> #include <algorithm> #include <Rtypes.h> #include "DataFormats/CaloRecHit/interface/CaloCluster.h" #include "DataFormats/CaloRecHit/interface/CaloClusterFwd.h" #include "DataFormats/Math/interface/Point3D.h" #include "DataFormats/ParticleFlowReco/interface/PFLayer.h" #include "DataFormats/ParticleFlowReco/interface/PFRecHit.h" #include "DataFormats/ParticleFlowReco/interface/PFRecHitFraction.h" #include "Math/GenVector/PositionVector3D.h" class PFClusterAlgo; namespace reco { /**\class PFCluster \brief Particle flow cluster, see clustering algorithm in PFClusterAlgo A particle flow cluster is defined by its energy and position, which are calculated from a vector of PFRecHitFraction. This calculation is performed in PFClusterAlgo. \todo Clean up this class to a common base (talk to Paolo Meridiani) the extra internal stuff (like the vector of PFRecHitFraction's) could be moved to a PFClusterExtra. \todo Now that PFRecHitFraction's hold a reference to the PFRecHit's, put back the calculation of energy and position to PFCluster. \todo Add an operator+= \author <NAME> \date July 2006 */ class PFCluster : public CaloCluster { public: typedef std::vector<std::pair<CaloClusterPtr::key_type, edm::Ptr<PFCluster>>> EEtoPSAssociation; // Next typedef uses double in ROOT 6 rather than Double32_t due to a bug in ROOT 5, // which otherwise would make ROOT5 files unreadable in ROOT6. This does not increase // the size on disk, because due to the bug, double was actually stored on disk in ROOT 5. typedef ROOT::Math::PositionVector3D<ROOT::Math::CylindricalEta3D<double>> REPPoint; PFCluster() : CaloCluster(CaloCluster::particleFlow), time_(-99.0), layer_(PFLayer::NONE) {} /// constructor PFCluster(PFLayer::Layer layer, double energy, double x, double y, double z); /// resets clusters parameters void reset(); /// reset only hits and fractions void resetHitsAndFractions(); /// add a given fraction of the rechit void addRecHitFraction(const reco::PFRecHitFraction& frac); /// vector of rechit fractions const std::vector<reco::PFRecHitFraction>& recHitFractions() const { return rechits_; } /// set layer void setLayer(PFLayer::Layer layer); /// cluster layer, see PFLayer.h in this directory PFLayer::Layer layer() const; /// cluster energy double energy() const { return energy_; } /// \return cluster time float time() const { return time_; } /// \return the timing uncertainty float timeError() const { return timeError_; } /// cluster depth double depth() const { return depth_; } void setTime(float time, float timeError = 0) { time_ = time; timeError_ = timeError; } void setTimeError(float timeError) { timeError_ = timeError; } void setDepth(double depth) { depth_ = depth; } /// cluster position: rho, eta, phi const REPPoint& positionREP() const { return posrep_; } /// computes posrep_ once and for all void calculatePositionREP() { posrep_.SetCoordinates(position_.Rho(), position_.Eta(), position_.Phi()); } /// \todo move to PFClusterTools static double getDepthCorrection(double energy, bool isBelowPS = false, bool isHadron = false); PFCluster& operator=(const PFCluster&); /// some classes to make this fit into a template footprint /// for RecoPFClusterRefCandidate so we can make jets and MET /// out of PFClusters. /// dummy charge double charge() const { return 0; } /// transverse momentum, massless approximation double pt() const { return (energy() * sin(position_.theta())); } /// angle double theta() const { return position_.theta(); } /// dummy vertex access math::XYZPoint const& vertex() const { return dummyVtx_; } double vx() const { return vertex().x(); } double vy() const { return vertex().y(); } double vz() const { return vertex().z(); } #if !defined(__CINT__) && !defined(__MAKECINT__) && !defined(__REFLEX__) template <typename pruner> void pruneUsing(pruner prune) { // remove_if+erase algo applied to both vectors... auto iter = std::find_if_not(rechits_.begin(), rechits_.end(), prune); if (iter == rechits_.end()) return; auto first = iter - rechits_.begin(); for (auto i = first; ++i < int(rechits_.size());) { if (prune(rechits_[i])) { rechits_[first] = std::move(rechits_[i]); hitsAndFractions_[first] = std::move(hitsAndFractions_[i]); ++first; } } rechits_.erase(rechits_.begin() + first, rechits_.end()); hitsAndFractions_.erase(hitsAndFractions_.begin() + first, hitsAndFractions_.end()); } #endif private: /// vector of rechit fractions (transient) std::vector<reco::PFRecHitFraction> rechits_; /// cluster position: rho, eta, phi (transient) REPPoint posrep_; /// Michalis: add timing and depth information float time_, timeError_; double depth_; /// transient layer PFLayer::Layer layer_; /// depth corrections static const constexpr double depthCorA_ = 0.89; static const constexpr double depthCorB_ = 7.3; static const constexpr double depthCorAp_ = 0.89; static const constexpr double depthCorBp_ = 4.0; static const math::XYZPoint dummyVtx_; }; std::ostream& operator<<(std::ostream& out, const PFCluster& cluster); } // namespace reco #endif // DataFormats_ParticleFlowReco_PFCluster_h
2,051
453
#include "util.h" #include <std/std.h> #include <std/memory.h> #include <kernel/util/paging/paging.h> #include <kernel/vmm/vmm.h> #include <kernel/boot_info.h> void move_stack(void* new_stack_start, uint32_t size) { Deprecated(); /* //allocate space for new stack printf("move_stack() mapping region 0x%08x to 0x%08x\n", new_stack_start - size, new_stack_start); //alloc 1 extra page at the top of the stack so if you don't page fault if you //access the very top of the stack vmm_map_region(vmm_active_pdir(), (uint32_t*)(new_stack_start - size), size + PAGING_PAGE_SIZE, PAGE_PRESENT_FLAG|PAGE_WRITE_FLAG); //flush TLB by reading and writing page directory address again printf_dbg("flushing TLB"); uint32_t pd_addr; asm volatile("mov %%cr3, %0" : "=r" (pd_addr)); asm volatile("mov %0, %%cr3" : : "r" (pd_addr)); //old ESP and EBP uint32_t old_sp; asm volatile("mov %%esp, %0" : "=r" (old_sp)); uint32_t old_bp; asm volatile("mov %%ebp, %0" : "=r" (old_bp)); //offset to add to old stack addresses to get new stack address uint32_t boot_esp = boot_info_get()->boot_stack_top_phys; uint32_t offset = (uint32_t)new_stack_start - boot_esp; //new esp and ebp uint32_t new_sp = old_sp + offset; uint32_t new_bp = old_bp + offset; printf_dbg("new ESP %x, new EBP %x", new_sp, new_bp); //copy stack! printf("copying stack data!\n"); memcpy((void*)new_sp, (void*)old_sp, boot_esp - old_sp); //backtrace through original stack, copying new values into new stack for (uint32_t i = (uint32_t)new_sp; i > (uint32_t)new_sp - size; i -= sizeof(uint32_t)) { uint32_t tmp = *(uint32_t*)i; //if value of tmp is inside range of old stack, //assume it's a base pointer and remap it //TODO keep in mind this will remap ANY value in this range, //whether it's a base pointer or not if ((old_sp < tmp) && (tmp < boot_esp)) { tmp = tmp + offset; uint32_t* tmp2 = (uint32_t*)i; *tmp2 = tmp; } } //change stacks asm volatile("mov %0, %%esp" : : "r" (new_sp)); asm volatile("mov %0, %%ebp" : : "r" (new_bp)); */ }
845
440
/*========================== begin_copyright_notice ============================ Copyright (C) 2020-2021 Intel Corporation SPDX-License-Identifier: MIT ============================= end_copyright_notice ===========================*/ #include "wa_def.h" #include "iadls_rev_id.h" #define ADLS_PCH_A0_REV_ID 0 //******************* Main Wa Initializer for Device Id ******************** // Initialize COMMON/DESKTOP/MOBILE WA using PLATFORM_STEP_APPLICABLE() macro. void InitAdlsSwWaTable(PWA_TABLE pWaTable, PSKU_FEATURE_TABLE pSkuTable, PWA_INIT_PARAM pWaParam) { int StepId_ADLS = (int)pWaParam->usRevId; int PchStepId_Adls = (int)pWaParam->usRevId_PCH; #ifdef __KCH // compilation issue with UTF: KCHASSERT(NULL != pWaParam); #endif //================================================================================================================= // // ADLS SW WA for all platforms // //================================================================================================================= SI_WA_ENABLE( WaMixModeSelInstDstNotPacked, "No HWBugLink provided", "No HWSightingLink provided", PLATFORM_ALL, SI_WA_FOR_EVER); } #ifdef __KCH void InitAdlsHASWaTable(PHW_DEVICE_EXTENSION pKchContext, PWA_TABLE pWaTable, PSKU_FEATURE_TABLE pSkuTable, PWA_INIT_PARAM pWaParam) { } #endif // __KCH
487
32,544
package com.baeldung.spring.patterns.proxy; import org.springframework.beans.factory.annotation.Autowired; import org.springframework.stereotype.Service; import org.springframework.transaction.annotation.Transactional; @Service public class BookManager { @Autowired private BookRepository repository; @Transactional public Book create(String author) { System.out.println(repository.getClass().getName()); return repository.create(author); } }
165
1,738
<gh_stars>1000+ #! /usr/bin/env python # encoding: utf-8 # harald at klimachs.de import re from waflib import Utils,Errors from waflib.Tools import fc,fc_config,fc_scan from waflib.Configure import conf from waflib.Tools.compiler_fc import fc_compiler fc_compiler['aix'].insert(0, 'fc_xlf') @conf def find_xlf(conf): """Find the xlf program (will look in the environment variable 'FC')""" fc = conf.find_program(['xlf2003_r', 'xlf2003', 'xlf95_r', 'xlf95', 'xlf90_r', 'xlf90', 'xlf_r', 'xlf'], var='FC') fc = conf.cmd_to_list(fc) conf.get_xlf_version(fc) conf.env.FC_NAME='XLF' @conf def xlf_flags(conf): v = conf.env v['FCDEFINES_ST'] = '-WF,-D%s' v['FCFLAGS_fcshlib'] = ['-qpic=small'] v['FCFLAGS_DEBUG'] = ['-qhalt=w'] v['LINKFLAGS_fcshlib'] = ['-Wl,-shared'] @conf def xlf_modifier_platform(conf): dest_os = conf.env['DEST_OS'] or Utils.unversioned_sys_platform() xlf_modifier_func = getattr(conf, 'xlf_modifier_' + dest_os, None) if xlf_modifier_func: xlf_modifier_func() @conf def get_xlf_version(conf, fc): """Get the compiler version""" cmd = fc + ['-qversion'] try: out, err = conf.cmd_and_log(cmd, output=0) except Errors.WafError: conf.fatal('Could not find xlf %r' % cmd) for v in (r"IBM XL Fortran.* V(?P<major>\d*)\.(?P<minor>\d*)",): version_re = re.compile(v, re.I).search match = version_re(out or err) if match: k = match.groupdict() conf.env['FC_VERSION'] = (k['major'], k['minor']) break else: conf.fatal('Could not determine the XLF version.') def configure(conf): conf.find_xlf() conf.find_ar() conf.fc_flags() conf.fc_add_flags() conf.xlf_flags() conf.xlf_modifier_platform()
738
4,303
<reponame>OAID/Halide<gh_stars>1000+ #include "runtime_internal.h" extern "C" int sched_yield(); extern "C" WEAK void halide_thread_yield() { sched_yield(); }
71
936
/* * Copyright 2020, Yahoo Inc. * Licensed under the Apache License, Version 2.0 * See LICENSE file in project root for terms. */ package com.yahoo.elide.datastores.aggregation.annotation.metricformula; import static com.yahoo.elide.datastores.aggregation.annotation.dimensionformula.DimensionFormulaTest.DUMMY_CONNECTION; import static org.junit.jupiter.api.Assertions.assertThrows; import static org.junit.jupiter.api.Assertions.assertTrue; import com.yahoo.elide.core.type.ClassType; import com.yahoo.elide.core.utils.DefaultClassScanner; import com.yahoo.elide.datastores.aggregation.metadata.MetaDataStore; import com.yahoo.elide.datastores.aggregation.queryengines.sql.SQLQueryEngine; import com.google.common.collect.Sets; import org.junit.jupiter.api.Test; public class MetricFormulaTest { @Test public void testReferenceLoop() { MetaDataStore metaDataStore = new MetaDataStore(DefaultClassScanner.getInstance(), Sets.newHashSet(ClassType.of(MeasureLoop.class)), true); IllegalArgumentException exception = assertThrows( IllegalArgumentException.class, () -> new SQLQueryEngine(metaDataStore, (unused) -> DUMMY_CONNECTION)); assertTrue(exception.getMessage().startsWith("Formula reference loop found:")); } }
455
1,333
<gh_stars>1000+ /* * Copyright 2014-2018 the original author or authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.codecentric.boot.admin.server.services.endpoints; import de.codecentric.boot.admin.server.domain.entities.Instance; import de.codecentric.boot.admin.server.domain.values.Endpoints; import de.codecentric.boot.admin.server.domain.values.InstanceId; import reactor.core.publisher.Mono; import reactor.test.StepVerifier; import org.junit.Test; import static org.assertj.core.api.Assertions.assertThatThrownBy; public class ChainingStrategyTest { @Test public void invariants() { assertThatThrownBy(() -> { new ChainingStrategy((EndpointDetectionStrategy[]) null); }).isInstanceOf(IllegalArgumentException.class).hasMessage("'delegates' must not be null."); assertThatThrownBy(() -> { new ChainingStrategy((EndpointDetectionStrategy) null); }).isInstanceOf(IllegalArgumentException.class).hasMessage("'delegates' must not contain null."); } @Test public void should_chain_on_empty() { //given Instance instance = Instance.create(InstanceId.of("id")); ChainingStrategy strategy = new ChainingStrategy((a) -> Mono.empty(), (a) -> Mono.empty(), (a) -> Mono.just(Endpoints.single("id", "path"))); //when/then StepVerifier.create(strategy.detectEndpoints(instance)) .expectNext(Endpoints.single("id", "path")) .verifyComplete(); } @Test public void should_return_empty_endpoints_when_all_empty() { //given Instance instance = Instance.create(InstanceId.of("id")); ChainingStrategy strategy = new ChainingStrategy((a) -> Mono.empty()); //when/then StepVerifier.create(strategy.detectEndpoints(instance)).expectNext(Endpoints.empty()).verifyComplete(); } }
865
438
import numpy as np import matplotlib as mpl import matplotlib.pyplot as plt import seaborn as sns np.random.seed(sum(map(ord, "aesthetics"))) # sns.set_style("white") x = [650, 1625, 3250, 4875, 6500] y = [38.04, 50.28, 54.76, 56.72, 61.20] y2 = [20.92, 31.08, 35.70, 40.16, 41.76] plt.plot(x, y, marker='o', markersize=12, linewidth=4) plt.plot(x, y2, marker='o', markersize=12, linewidth=4) plt.xlabel('Number of Training Examples', fontsize=24) plt.ylabel('Model Accuracy (%)', fontsize=26) plt.title('DeepRegex Performance vs. Data Size', fontsize=32) plt.legend(['DFA-Equal', 'String-Equal'], loc='upper left') leg = plt.gca().get_legend() ltext = leg.get_texts() plt.setp(ltext, fontsize=18) plt.tick_params(labelsize=16) plt.tick_params(labelsize=16) plt.show()
339
1,041
package org.tests.model.history; import org.tests.model.draftable.BaseDomain; import javax.persistence.Entity; import javax.persistence.ManyToMany; import java.util.List; @Entity public class HeDoc extends BaseDomain { String name; @ManyToMany(mappedBy = "docs") List<HeLink> links; public HeDoc(String name) { this.name = name; } public String getName() { return name; } public void setName(String name) { this.name = name; } public List<HeLink> getLinks() { return links; } public void setLinks(List<HeLink> links) { this.links = links; } }
216
1,248
<gh_stars>1000+ # -*- coding: utf-8 -*- # Generated by Django 1.9.2 on 2016-04-23 09:44 from __future__ import unicode_literals from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('pretixbase', '0020_auto_20160421_1943'), ('pretixbase', '0021_auto_20160418_2117'), ] operations = [ ]
151
1,738
<filename>dev/Gems/Multiplayer/Code/Source/MultiplayerCVars.h /* * All or portions of this file Copyright (c) Amazon.com, Inc. or its affiliates or * its licensors. * * For complete copyright and license terms please see the LICENSE at the root of this * distribution (the "License"). All use of this software is governed by the License, * or, if provided, by the license below or the license accompanying this file. Do not * remove or modify any license notices. This file is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * */ #ifndef INCLUDE_MULTIPLAYERCVARS_HEADER #define INCLUDE_MULTIPLAYERCVARS_HEADER #include <MultiplayerGameLiftClient.h> #include <INetwork.h> #include <CertificateManager/ICertificateManagerGem.h> #include <GridMate/Session/Session.h> struct IConsole; struct IConsoleCmdArgs; namespace GridMate { class GridSearch; } namespace CertificateManager { class FileDataSource; } namespace Multiplayer { /*! * GridMate-specific network cvars. */ class MultiplayerCVars : public GridMate::SessionEventBus::Handler { public: MultiplayerCVars(); ~MultiplayerCVars(); void RegisterCVars(); void UnregisterCVars(); void PostInitRegistration(); private: //! Host a session (LAN). static void MPHostLANCmd(IConsoleCmdArgs* args); //! Attempt to join an existing session (LAN). static void MPJoinLANCmd(IConsoleCmdArgs* args); #if defined(BUILD_GAMELIFT_CLIENT) //! Attempt to host a session on GameLift and join it. static void MPHostGameLiftCmd(IConsoleCmdArgs* args); //! Attempt to join an existing GameLift session. static void MPJoinGameLiftCmd(IConsoleCmdArgs* args); static void StopGameLiftClient(IConsoleCmdArgs* args); static void MPMatchmakingGameLiftCmd(IConsoleCmdArgs* args); #endif //! Shut down current server or client session. static void MPDisconnectCmd(IConsoleCmdArgs* args); private: void OnGridSearchComplete(GridMate::GridSearch* gridSearch) override; #if defined(NET_SUPPORT_SECURE_SOCKET_DRIVER) static void OnPrivateKeyChanged(ICVar* cvar); static void OnCertificateChanged(ICVar* cvar); static void OnCAChanged(ICVar* cvar); static void CreateFileDataSource(); #endif bool m_autoJoin; GridMate::GridSearch* m_search; #if defined(BUILD_GAMELIFT_CLIENT) MultiplayerGameLiftClient m_gameLift; #endif static MultiplayerCVars* s_instance; }; } // namespace GridMate #endif // INCLUDE_NETWORKGRIDMATECVARS_HEADER
991
2,625
/*------------------------------------------------------------------------- * * cmsketch.h * Interface for Count-Min Sketch support * * Copyright (c) 2018, PipelineDB, Inc. * *------------------------------------------------------------------------- */ #ifndef PIPELINE_CMSKETCH_H #define PIPELINE_CMSKETCH_H #include "c.h" /* * Unfortunately we can't do top-K with Count-Min Sketch for continuous * queries because it will require us to store O(n) values. This isn't a * problem in the case when Count-Min Sketch structures are not being * merged in which case only O(1) values need to be stored. */ typedef struct CountMinSketch { uint32 vl_len_; uint64_t count; uint32_t d; uint32_t w; uint32_t table[1]; } CountMinSketch; extern CountMinSketch *CountMinSketchCreateWithDAndW(uint32_t d, uint32_t w); extern CountMinSketch *CountMinSketchCreateWithEpsAndP(float8 epsilon, float8 p); extern CountMinSketch *CountMinSketchCreate(void); extern void CountMinSketchDestroy(CountMinSketch *cms); extern CountMinSketch *CountMinSketchCopy(CountMinSketch *cms); extern void CountMinSketchAdd(CountMinSketch *cms, void *key, Size size, uint32_t count); extern uint32_t CountMinSketchEstimateFrequency(CountMinSketch *cms, void *key, Size size); extern float8 CountMinSketchEstimateNormFrequency(CountMinSketch *cms, void *key, Size size); extern uint64_t CountMinSketchTotal(CountMinSketch *cms); extern CountMinSketch *CountMinSketchMerge(CountMinSketch *result, CountMinSketch* incoming); extern Size CountMinSketchSize(CountMinSketch *cms); #endif
526
997
<gh_stars>100-1000 #include <stdint.h> #include <string.h> #include "address.h" #include "hash.h" #include "params.h" #include "utils.h" #include "fips202.h" /* For SHAKE256, there is no immediate reason to initialize at the start, so this function is an empty operation. */ void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_initialize_hash_function( hash_state *hash_state_seeded, // NOLINT(readability-non-const-parameter) const unsigned char *pub_seed, const unsigned char *sk_seed) { (void)hash_state_seeded; /* Suppress an 'unused parameter' warning. */ (void)pub_seed; /* Suppress an 'unused parameter' warning. */ (void)sk_seed; /* Suppress an 'unused parameter' warning. */ } /* This is not necessary for SHAKE256, so we don't do anything */ void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_destroy_hash_function( hash_state *hash_state_seeded) { // NOLINT(readability-non-const-parameter) (void)hash_state_seeded; } /* * Computes PRF(key, addr), given a secret key of PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N bytes and an address */ void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_prf_addr( unsigned char *out, const unsigned char *key, const uint32_t addr[8], const hash_state *hash_state_seeded) { unsigned char buf[PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_ADDR_BYTES]; memcpy(buf, key, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N); PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_addr_to_bytes(buf + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N, addr); shake256(out, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N, buf, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_ADDR_BYTES); (void)hash_state_seeded; /* Prevent unused parameter warning. */ } /** * Computes the message-dependent randomness R, using a secret seed and an * optional randomization value as well as the message. */ void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_gen_message_random( unsigned char *R, const unsigned char *sk_prf, const unsigned char *optrand, const unsigned char *m, size_t mlen, const hash_state *hash_state_seeded) { shake256incctx state; shake256_inc_init(&state); shake256_inc_absorb(&state, sk_prf, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N); shake256_inc_absorb(&state, optrand, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N); shake256_inc_absorb(&state, m, mlen); shake256_inc_finalize(&state); shake256_inc_squeeze(R, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N, &state); shake256_inc_ctx_release(&state); (void)hash_state_seeded; /* Prevent unused parameter warning. */ } /** * Computes the message hash using R, the public key, and the message. * Outputs the message digest and the index of the leaf. The index is split in * the tree index and the leaf index, for convenient copying to an address. */ void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_hash_message( unsigned char *digest, uint64_t *tree, uint32_t *leaf_idx, const unsigned char *R, const unsigned char *pk, const unsigned char *m, size_t mlen, const hash_state *hash_state_seeded) { #define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BITS (PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_HEIGHT * (PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_D - 1)) #define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BYTES ((PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BITS + 7) / 8) #define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BITS PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_HEIGHT #define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BYTES ((PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BITS + 7) / 8) #define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_DGST_BYTES (PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_FORS_MSG_BYTES + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BYTES + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BYTES) unsigned char buf[PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_DGST_BYTES]; unsigned char *bufp = buf; shake256incctx state; shake256_inc_init(&state); shake256_inc_absorb(&state, R, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N); shake256_inc_absorb(&state, pk, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_PK_BYTES); shake256_inc_absorb(&state, m, mlen); shake256_inc_finalize(&state); shake256_inc_squeeze(buf, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_DGST_BYTES, &state); shake256_inc_ctx_release(&state); memcpy(digest, bufp, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_FORS_MSG_BYTES); bufp += PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_FORS_MSG_BYTES; *tree = PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_bytes_to_ull( bufp, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BYTES); *tree &= (~(uint64_t)0) >> (64 - PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BITS); bufp += PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BYTES; *leaf_idx = (uint32_t)PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_bytes_to_ull( bufp, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BYTES); *leaf_idx &= (~(uint32_t)0) >> (32 - PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BITS); (void)hash_state_seeded; /* Prevent unused parameter warning. */ }
2,282
1,582
# -*- coding: utf-8 -*- import json from vilya.libs import api_errors from vilya.models.project import CodeDoubanProject from vilya.views.api.utils import RestAPIUI, api_require_login, jsonize from vilya.views.api.repos.product import ProductUI from vilya.views.api.repos.summary import SummaryUI from vilya.views.api.repos.intern import InternUI from vilya.views.api.repos.default_branch import DefaultBranchUI from vilya.views.api.repos.commits import CommitsUI from vilya.views.api.repos.post_receive import PostReceiveUI from vilya.views.api.repos.git2svn import GIT2SVNUI from vilya.views.api.repos.svn2git import SVN2GITUI from vilya.views.api.repos.pulls import PullsUI from vilya.views.api.repos.issues import IssuesUI from vilya.views.api.repos.contents import ContentsUI from vilya.views.api.repos.push import PushUI from vilya.views.api.repos.watchers import WatchersUI _q_exports = [] def _q_lookup(request, name): return RepositoryUI(name) def _q_access(request): request.response.set_content_type('application/json; charset=utf-8') class RepositoryUI(object): _q_exports = [ 'lang_stats', 'forks', 'pulls', 'summary', 'committers', 'name', 'owner', 'product', 'intern_banned', 'default_branch', 'commits', 'post_receive', 'svn2git', 'git2svn', 'issues', 'contents', 'can_push', 'watchers' ] def __init__(self, name): self.name = name self.repo = CodeDoubanProject.get_by_name(self.name) def __call__(self, request): return self._q_index(request) @jsonize def _q_index(self, request): if not self.repo: raise api_errors.NotFoundError("repo") return {} def _q_access(self, request): self.method = request.method def _q_lookup(self, request, part): name = "%s/%s" % (self.name, part) if not CodeDoubanProject.exists(name): raise api_errors.NotFoundError("repo") return RepositoryUI(name) @jsonize def lang_stats(self, request): if not self.repo: raise api_errors.NotFoundError if self.method == 'POST': language = request.get_form_var('language', '') languages = request.get_form_var('languages', '[]') try: languages = json.loads(languages) except ValueError: raise api_errors.NotJSONError self.repo.language = language self.repo.languages = languages return {} else: return dict(language=self.repo.language, languages=self.repo.languages) @property def forks(self): return ForksUI(self.repo) @property def pulls(self): return PullsUI(self.repo) @property def product(self): return ProductUI(self.repo) @property def summary(self): return SummaryUI(self.repo) @property def intern_banned(self): return InternUI(self.repo) @property def can_push(self): return PushUI(self.repo) @property def default_branch(self): return DefaultBranchUI(self.repo) @property def commits(self): return CommitsUI(self.repo) @property def post_receive(self): return PostReceiveUI(self.repo) @property def svn2git(self): return SVN2GITUI(self.repo) @property def git2svn(self): return GIT2SVNUI(self.repo) @property def issues(self): return IssuesUI(self.repo) @property def contents(self): return ContentsUI(self.repo) @property def watchers(self): return WatchersUI(self.repo) class ForksUI(RestAPIUI): _q_exports = [] _q_methods = ['get', 'post'] def __init__(self, repo): self.repo = repo @api_require_login def post(self, request): repo = self.repo fork_repo = repo.new_fork(self.user.name) if not fork_repo: # FIXME: repository exists return [] return fork_repo.as_dict() def get(self, request): fork_repos = self.repo.get_forked_projects() return [project.get_info(without_commits=True) for project in fork_repos]
1,899
471
<reponame>madanagopaltcomcast/pxCore ///////////////////////////////////////////////////////////////////////////// // Name: wx/gtk1/checkbox.h // Purpose: // Author: <NAME> // Copyright: (c) 1998 <NAME> // Licence: wxWindows licence ///////////////////////////////////////////////////////////////////////////// #ifndef __GTKCHECKBOXH__ #define __GTKCHECKBOXH__ // ---------------------------------------------------------------------------- // wxCheckBox // ---------------------------------------------------------------------------- class WXDLLIMPEXP_CORE wxCheckBox : public wxCheckBoxBase { public: wxCheckBox(); wxCheckBox( wxWindow *parent, wxWindowID id, const wxString& label, const wxPoint& pos = wxDefaultPosition, const wxSize& size = wxDefaultSize, long style = 0, const wxValidator& validator = wxDefaultValidator, const wxString& name = wxCheckBoxNameStr) { Create(parent, id, label, pos, size, style, validator, name); } bool Create(wxWindow *parent, wxWindowID id, const wxString& label, const wxPoint& pos = wxDefaultPosition, const wxSize& size = wxDefaultSize, long style = 0, const wxValidator& validator = wxDefaultValidator, const wxString& name = wxCheckBoxNameStr ); void SetValue( bool state ); bool GetValue() const; virtual void SetLabel( const wxString& label ); virtual bool Enable( bool enable = TRUE ); static wxVisualAttributes GetClassDefaultAttributes(wxWindowVariant variant = wxWINDOW_VARIANT_NORMAL); // implementation // -------------- void DoApplyWidgetStyle(GtkRcStyle *style); bool IsOwnGtkWindow( GdkWindow *window ); void OnInternalIdle(); GtkWidget *m_widgetCheckbox; GtkWidget *m_widgetLabel; bool m_blockEvent; protected: virtual wxSize DoGetBestSize() const; private: DECLARE_DYNAMIC_CLASS(wxCheckBox) }; #endif // __GTKCHECKBOXH__
777
4,036
<reponame>vadi2/codeql<filename>cpp/ql/test/library-tests/proxy_class/proxy_class.cpp struct Base {}; template <typename T> struct Derived {friend class T;}; int main() { return sizeof(Derived<Base>); } // semmle-extractor-options: --microsoft
87
5,964
#ifndef V8_TORQUE_TYPED_ARRAY_SUBARRAY_FROM_DSL_BASE_H__ #define V8_TORQUE_TYPED_ARRAY_SUBARRAY_FROM_DSL_BASE_H__ #include "src/compiler/code-assembler.h" #include "src/code-stub-assembler.h" #include "src/utils.h" #include "torque-generated/class-definitions-from-dsl.h" namespace v8 { namespace internal { class TypedArraySubarrayBuiltinsFromDSLAssembler { public: explicit TypedArraySubarrayBuiltinsFromDSLAssembler(compiler::CodeAssemblerState* state) : state_(state), ca_(state) { USE(state_, ca_); } private: compiler::CodeAssemblerState* const state_; compiler::CodeAssembler ca_; }; } // namespace internal } // namespace v8 #endif // V8_TORQUE_TYPED_ARRAY_SUBARRAY_FROM_DSL_BASE_H__
291
2,293
_G='r' _F='w' _E=False _D=True _C='strict' _B='utf-8' _A=None import csv,json,sys,warnings from pathlib import Path import dynaconf.vendor.ruamel.yaml as yaml from dynaconf.vendor.box.exceptions import BoxError,BoxWarning from dynaconf.vendor import toml BOX_PARAMETERS='default_box','default_box_attr','conversion_box','frozen_box','camel_killer_box','box_safe_prefix','box_duplicates','ordered_box','default_box_none_transform','box_dots','modify_tuples_box','box_intact_types','box_recast' def _exists(filename,create=_E): A=filename;B=Path(A) if create: try:B.touch(exist_ok=_D) except OSError as C:raise BoxError(f"Could not create file {A} - {C}") else:return if not B.exists():raise BoxError(f'File "{A}" does not exist') if not B.is_file():raise BoxError(f"{A} is not a file") def _to_json(obj,filename=_A,encoding=_B,errors=_C,**C): A=filename;B=json.dumps(obj,ensure_ascii=_E,**C) if A: _exists(A,create=_D) with open(A,_F,encoding=encoding,errors=errors)as D:D.write(B if sys.version_info>=(3,0)else B.decode(_B)) else:return B def _from_json(json_string=_A,filename=_A,encoding=_B,errors=_C,multiline=_E,**B): D=json_string;A=filename if A: _exists(A) with open(A,_G,encoding=encoding,errors=errors)as E: if multiline:C=[json.loads(A.strip(),**B)for A in E if A.strip()and not A.strip().startswith('#')] else:C=json.load(E,**B) elif D:C=json.loads(D,**B) else:raise BoxError('from_json requires a string or filename') return C def _to_yaml(obj,filename=_A,default_flow_style=_E,encoding=_B,errors=_C,**C): B=default_flow_style;A=filename if A: _exists(A,create=_D) with open(A,_F,encoding=encoding,errors=errors)as D:yaml.dump(obj,stream=D,default_flow_style=B,**C) else:return yaml.dump(obj,default_flow_style=B,**C) def _from_yaml(yaml_string=_A,filename=_A,encoding=_B,errors=_C,**A): F='Loader';C=yaml_string;B=filename if F not in A:A[F]=yaml.SafeLoader if B: _exists(B) with open(B,_G,encoding=encoding,errors=errors)as E:D=yaml.load(E,**A) elif C:D=yaml.load(C,**A) else:raise BoxError('from_yaml requires a string or filename') return D def _to_toml(obj,filename=_A,encoding=_B,errors=_C): A=filename if A: _exists(A,create=_D) with open(A,_F,encoding=encoding,errors=errors)as B:toml.dump(obj,B) else:return toml.dumps(obj) def _from_toml(toml_string=_A,filename=_A,encoding=_B,errors=_C): B=toml_string;A=filename if A: _exists(A) with open(A,_G,encoding=encoding,errors=errors)as D:C=toml.load(D) elif B:C=toml.loads(B) else:raise BoxError('from_toml requires a string or filename') return C def _to_csv(box_list,filename,encoding=_B,errors=_C): B=filename;A=box_list;C=list(A[0].keys()) for E in A: if list(E.keys())!=C:raise BoxError('BoxList must contain the same dictionary structure for every item to convert to csv') if B: _exists(B,create=_D) with open(B,_F,encoding=encoding,errors=errors,newline='')as F: D=csv.DictWriter(F,fieldnames=C);D.writeheader() for G in A:D.writerow(G) def _from_csv(filename,encoding=_B,errors=_C): A=filename;_exists(A) with open(A,_G,encoding=encoding,errors=errors,newline='')as B:C=csv.DictReader(B);return[A for A in C]
1,410
471
<reponame>dimagilg/commcare-hq class NoAccountException(Exception): """ Raised when trying to access the account of someone without one """ pass class InvalidMobileWorkerRequest(Exception): pass class IllegalAccountConfirmation(Exception): pass
87
2,231
#!/usr/bin/env python # Copyright 2020 The Defold Foundation # Licensed under the Defold License version 1.0 (the "License"); you may not use # this file except in compliance with the License. # # You may obtain a copy of the License, together with FAQs at # https://www.defold.com/license # # Unless required by applicable law or agreed to in writing, software distributed # under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR # CONDITIONS OF ANY KIND, either express or implied. See the License for the # specific language governing permissions and limitations under the License. import Tkinter import glob, stat, os, urllib root_path = "content" state = {} def UpdateState(): changed = [] for path, dirs, files in os.walk(root_path): (root, sep, sub_path) = path.partition(os.path.sep) for f in files: name = os.path.join(path, f) mtime = os.stat(name)[stat.ST_MTIME] oldmtime = state.get(name, 0) if mtime != oldmtime: changed.append(os.path.join(sub_path, f + "c")) state[name] = mtime return changed root = Tkinter.Tk() root.minsize(300, 200) label = Tkinter.Label() label.config(text = "Waiting for resources to change") label.pack() UpdateState() def tick(): label.config(text = "Waiting for resources to change") changed = UpdateState() if len(changed) > 0: label.config(text = "Compiling...") label.update() result = os.system('waf --skip-tests') if result == 0: try: for f in changed: print("Reloading: " + f) request = urllib.urlopen("http://localhost:8001/reload/" + f) label.config(text = "Success!", foreground = "#080") except: label.config(text = "No connection!", foreground = "#880") else: label.config(text = "Failure!", foreground = "#a00") label.after(1000, tick) tick() label.mainloop()
818
603
"""CFFI backend (for PyPy)""" # Copyright (C) PyZMQ Developers # Distributed under the terms of the Modified BSD License. from zmq.backend.cffi import ( constants, error, message, context, socket, _poll, devices, utils, ) from ._cffi import ffi, lib as C def zmq_version_info(): """Get libzmq version as tuple of ints""" major = ffi.new('int*') minor = ffi.new('int*') patch = ffi.new('int*') C.zmq_version(major, minor, patch) return (int(major[0]), int(minor[0]), int(patch[0])) __all__ = ["zmq_version_info"] for submod in (constants, error, message, context, socket, _poll, devices, utils): __all__.extend(submod.__all__) from .constants import * from .error import * from .message import * from .context import * from .socket import * from .devices import * from ._poll import * from .utils import *
339
2,114
package io.searchbox.client.http; import com.google.gson.Gson; import io.searchbox.action.Action; import io.searchbox.client.AbstractJestClient; import io.searchbox.client.JestResult; import io.searchbox.client.JestResultHandler; import io.searchbox.client.config.ElasticsearchVersion; import io.searchbox.client.config.exception.CouldNotConnectException; import io.searchbox.client.http.apache.HttpDeleteWithEntity; import io.searchbox.client.http.apache.HttpGetWithEntity; import org.apache.http.Header; import org.apache.http.HttpEntityEnclosingRequest; import org.apache.http.HttpRequest; import org.apache.http.HttpResponse; import org.apache.http.StatusLine; import org.apache.http.client.config.RequestConfig; import org.apache.http.client.entity.EntityBuilder; import org.apache.http.client.methods.CloseableHttpResponse; import org.apache.http.client.methods.HttpHead; import org.apache.http.client.methods.HttpPost; import org.apache.http.client.methods.HttpPut; import org.apache.http.client.methods.HttpRequestBase; import org.apache.http.client.methods.HttpUriRequest; import org.apache.http.client.protocol.HttpClientContext; import org.apache.http.concurrent.FutureCallback; import org.apache.http.conn.HttpHostConnectException; import org.apache.http.entity.ContentType; import org.apache.http.impl.client.CloseableHttpClient; import org.apache.http.impl.nio.client.CloseableHttpAsyncClient; import org.apache.http.util.EntityUtils; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import java.io.IOException; import java.util.Map.Entry; import java.util.concurrent.Future; /** * @author <NAME> * @author <NAME> */ public class JestHttpClient extends AbstractJestClient { private final static Logger log = LoggerFactory.getLogger(JestHttpClient.class); protected ContentType requestContentType = ContentType.APPLICATION_JSON.withCharset("utf-8"); private CloseableHttpClient httpClient; private CloseableHttpAsyncClient asyncClient; private HttpClientContext httpClientContextTemplate; private ElasticsearchVersion elasticsearchVersion = ElasticsearchVersion.UNKNOWN; /** * @throws IOException in case of a problem or the connection was aborted during request, * or in case of a problem while reading the response stream * @throws CouldNotConnectException if an {@link HttpHostConnectException} is encountered */ @Override public <T extends JestResult> T execute(Action<T> clientRequest) throws IOException { return execute(clientRequest, null); } public <T extends JestResult> T execute(Action<T> clientRequest, RequestConfig requestConfig) throws IOException { HttpUriRequest request = prepareRequest(clientRequest, requestConfig); CloseableHttpResponse response = null; try { response = executeRequest(request); return deserializeResponse(response, request, clientRequest); } catch (HttpHostConnectException ex) { throw new CouldNotConnectException(ex.getHost().toURI(), ex); } finally { if (response != null) { try { response.close(); } catch (IOException ex) { log.error("Exception occurred while closing response stream.", ex); } } } } @Override public <T extends JestResult> void executeAsync(final Action<T> clientRequest, final JestResultHandler<? super T> resultHandler) { executeAsync(clientRequest, resultHandler, null); } public <T extends JestResult> void executeAsync(final Action<T> clientRequest, final JestResultHandler<? super T> resultHandler, final RequestConfig requestConfig) { synchronized (this) { if (!asyncClient.isRunning()) { asyncClient.start(); } } HttpUriRequest request = prepareRequest(clientRequest, requestConfig); executeAsyncRequest(clientRequest, resultHandler, request); } @Override public void shutdownClient() { try { close(); } catch (IOException e) { log.error("Exception occurred while shutting down the sync client.", e); } } @Override public void close() throws IOException { super.close(); asyncClient.close(); httpClient.close(); } protected <T extends JestResult> HttpUriRequest prepareRequest(final Action<T> clientRequest, final RequestConfig requestConfig) { String elasticSearchRestUrl = getRequestURL(getNextServer(), clientRequest.getURI(elasticsearchVersion)); HttpUriRequest request = constructHttpMethod(clientRequest.getRestMethodName(), elasticSearchRestUrl, clientRequest.getData(gson), requestConfig); log.debug("Request method={} url={}", clientRequest.getRestMethodName(), elasticSearchRestUrl); // add headers added to action for (Entry<String, Object> header : clientRequest.getHeaders().entrySet()) { request.addHeader(header.getKey(), header.getValue().toString()); } return request; } protected CloseableHttpResponse executeRequest(HttpUriRequest request) throws IOException { if (httpClientContextTemplate != null) { return httpClient.execute(request, createContextInstance()); } return httpClient.execute(request); } protected <T extends JestResult> Future<HttpResponse> executeAsyncRequest(Action<T> clientRequest, JestResultHandler<? super T> resultHandler, HttpUriRequest request) { if (httpClientContextTemplate != null) { return asyncClient.execute(request, createContextInstance(), new DefaultCallback<T>(clientRequest, request, resultHandler)); } return asyncClient.execute(request, new DefaultCallback<T>(clientRequest, request, resultHandler)); } protected HttpClientContext createContextInstance() { HttpClientContext context = HttpClientContext.create(); context.setCredentialsProvider(httpClientContextTemplate.getCredentialsProvider()); context.setAuthCache(httpClientContextTemplate.getAuthCache()); return context; } protected HttpUriRequest constructHttpMethod(String methodName, String url, String payload, RequestConfig requestConfig) { HttpUriRequest httpUriRequest = null; if (methodName.equalsIgnoreCase("POST")) { httpUriRequest = new HttpPost(url); log.debug("POST method created based on client request"); } else if (methodName.equalsIgnoreCase("PUT")) { httpUriRequest = new HttpPut(url); log.debug("PUT method created based on client request"); } else if (methodName.equalsIgnoreCase("DELETE")) { httpUriRequest = new HttpDeleteWithEntity(url); log.debug("DELETE method created based on client request"); } else if (methodName.equalsIgnoreCase("GET")) { httpUriRequest = new HttpGetWithEntity(url); log.debug("GET method created based on client request"); } else if (methodName.equalsIgnoreCase("HEAD")) { httpUriRequest = new HttpHead(url); log.debug("HEAD method created based on client request"); } if (httpUriRequest instanceof HttpRequestBase && requestConfig != null) { ((HttpRequestBase) httpUriRequest).setConfig(requestConfig); } if (httpUriRequest != null && httpUriRequest instanceof HttpEntityEnclosingRequest && payload != null) { EntityBuilder entityBuilder = EntityBuilder.create() .setText(payload) .setContentType(requestContentType); if (isRequestCompressionEnabled()) { entityBuilder.gzipCompress(); } ((HttpEntityEnclosingRequest) httpUriRequest).setEntity(entityBuilder.build()); } return httpUriRequest; } private <T extends JestResult> T deserializeResponse(HttpResponse response, final HttpRequest httpRequest, Action<T> clientRequest) throws IOException { StatusLine statusLine = response.getStatusLine(); try { return clientRequest.createNewElasticSearchResult( response.getEntity() == null ? null : EntityUtils.toString(response.getEntity()), statusLine.getStatusCode(), statusLine.getReasonPhrase(), gson ); } catch (com.google.gson.JsonSyntaxException e) { for (Header header : response.getHeaders("Content-Type")) { final String mimeType = header.getValue(); if (!mimeType.startsWith("application/json")) { // probably a proxy that responded in text/html final String message = "Request " + httpRequest.toString() + " yielded " + mimeType + ", should be json: " + statusLine.toString(); throw new IOException(message, e); } } throw e; } } public CloseableHttpClient getHttpClient() { return httpClient; } public void setHttpClient(CloseableHttpClient httpClient) { this.httpClient = httpClient; } public CloseableHttpAsyncClient getAsyncClient() { return asyncClient; } public void setAsyncClient(CloseableHttpAsyncClient asyncClient) { this.asyncClient = asyncClient; } public Gson getGson() { return gson; } public void setGson(Gson gson) { this.gson = gson; } public HttpClientContext getHttpClientContextTemplate() { return httpClientContextTemplate; } public void setHttpClientContextTemplate(HttpClientContext httpClientContext) { this.httpClientContextTemplate = httpClientContext; } public void setElasticsearchVersion(ElasticsearchVersion elasticsearchVersion) { this.elasticsearchVersion = elasticsearchVersion; } protected class DefaultCallback<T extends JestResult> implements FutureCallback<HttpResponse> { private final Action<T> clientRequest; private final HttpRequest request; private final JestResultHandler<? super T> resultHandler; public DefaultCallback(Action<T> clientRequest, final HttpRequest request, JestResultHandler<? super T> resultHandler) { this.clientRequest = clientRequest; this.request = request; this.resultHandler = resultHandler; } @Override public void completed(final HttpResponse response) { T jestResult = null; try { jestResult = deserializeResponse(response, request, clientRequest); } catch (Exception e) { failed(e); } catch (Throwable t) { failed(new Exception("Problem during request processing", t)); } if (jestResult != null) resultHandler.completed(jestResult); } @Override public void failed(final Exception ex) { log.error("Exception occurred during async execution.", ex); if (ex instanceof HttpHostConnectException) { String host = ((HttpHostConnectException) ex).getHost().toURI(); resultHandler.failed(new CouldNotConnectException(host, ex)); return; } resultHandler.failed(ex); } @Override public void cancelled() { log.warn("Async execution was cancelled; this is not expected to occur under normal operation."); } } }
4,401
679
<reponame>Grosskopf/openoffice /************************************************************** * * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. * *************************************************************/ #ifndef SC_EDITSRC_HXX #define SC_EDITSRC_HXX #include "address.hxx" #include <editeng/unoedsrc.hxx> #include <svl/lstner.hxx> #include <memory> class ScEditEngineDefaulter; class SvxEditEngineForwarder; class ScDocShell; class ScHeaderFooterContentObj; class ScCellTextData; class ScHeaderFooterTextData; class ScAccessibleTextData; class SdrObject; class ScHeaderFooterChangedHint : public SfxHint { sal_uInt16 nPart; public: TYPEINFO(); ScHeaderFooterChangedHint(sal_uInt16 nP); ~ScHeaderFooterChangedHint(); sal_uInt16 GetPart() const { return nPart; } }; // all ScSharedHeaderFooterEditSource objects for a single text share the same data class ScSharedHeaderFooterEditSource : public SvxEditSource { private: ScHeaderFooterTextData* pTextData; protected: ScHeaderFooterTextData* GetTextData() const { return pTextData; } // for ScHeaderFooterEditSource public: ScSharedHeaderFooterEditSource( ScHeaderFooterTextData* pData ); virtual ~ScSharedHeaderFooterEditSource(); // GetEditEngine is needed because the forwarder doesn't have field functions ScEditEngineDefaulter* GetEditEngine(); virtual SvxEditSource* Clone() const ; virtual SvxTextForwarder* GetTextForwarder(); virtual void UpdateData(); }; // ScHeaderFooterEditSource with local copy of ScHeaderFooterTextData is used by field objects class ScHeaderFooterEditSource : public ScSharedHeaderFooterEditSource { public: ScHeaderFooterEditSource( ScHeaderFooterContentObj* pContent, sal_uInt16 nP ); ScHeaderFooterEditSource( ScHeaderFooterContentObj& rContent, sal_uInt16 nP ); virtual ~ScHeaderFooterEditSource(); virtual SvxEditSource* Clone() const; }; // Data (incl. EditEngine) for cell EditSource is now shared in ScCellTextData class ScSharedCellEditSource : public SvxEditSource { private: ScCellTextData* pCellTextData; protected: ScCellTextData* GetCellTextData() const { return pCellTextData; } // for ScCellEditSource public: ScSharedCellEditSource( ScCellTextData* pData ); virtual ~ScSharedCellEditSource(); // GetEditEngine is needed because the forwarder doesn't have field functions ScEditEngineDefaulter* GetEditEngine(); virtual SvxEditSource* Clone() const; virtual SvxTextForwarder* GetTextForwarder(); virtual void UpdateData(); void SetDoUpdateData(sal_Bool bValue); sal_Bool IsDirty() const; }; // ScCellEditSource with local copy of ScCellTextData is used by ScCellFieldsObj, ScCellFieldObj class ScCellEditSource : public ScSharedCellEditSource { public: ScCellEditSource( ScDocShell* pDocSh, const ScAddress& rP ); virtual ~ScCellEditSource(); virtual SvxEditSource* Clone() const; }; class ScAnnotationEditSource : public SvxEditSource, public SfxListener { private: ScDocShell* pDocShell; ScAddress aCellPos; ScEditEngineDefaulter* pEditEngine; SvxEditEngineForwarder* pForwarder; sal_Bool bDataValid; SdrObject* GetCaptionObj(); public: ScAnnotationEditSource(ScDocShell* pDocSh, const ScAddress& rP); virtual ~ScAnnotationEditSource(); virtual SvxEditSource* Clone() const ; virtual SvxTextForwarder* GetTextForwarder(); virtual void UpdateData(); virtual void Notify( SfxBroadcaster& rBC, const SfxHint& rHint ); }; // EditSource with a shared forwarder for all children of one text object class ScSimpleEditSource : public SvxEditSource { private: SvxTextForwarder* pForwarder; public: ScSimpleEditSource( SvxTextForwarder* pForw ); virtual ~ScSimpleEditSource(); virtual SvxEditSource* Clone() const ; virtual SvxTextForwarder* GetTextForwarder(); virtual void UpdateData(); }; class ScAccessibilityEditSource : public SvxEditSource { private: ::std::auto_ptr < ScAccessibleTextData > mpAccessibleTextData; public: ScAccessibilityEditSource( ::std::auto_ptr < ScAccessibleTextData > pAccessibleCellTextData ); virtual ~ScAccessibilityEditSource(); virtual SvxEditSource* Clone() const; virtual SvxTextForwarder* GetTextForwarder(); virtual SvxViewForwarder* GetViewForwarder(); virtual SvxEditViewForwarder* GetEditViewForwarder( sal_Bool bCreate = sal_False ); virtual void UpdateData(); virtual SfxBroadcaster& GetBroadcaster() const; }; #endif
1,778
17,481
/* * Copyright (C) 2021 The Dagger Authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package dagger.hilt.android.internal; import android.app.Application; import android.content.Context; import android.content.ContextWrapper; /** * Utility methods for dealing with contexts. */ public final class Contexts { /** Finds the android Application from a context. */ public static Application getApplication(Context context) { if (context instanceof Application) { return (Application) context; } Context unwrapContext = context; while (unwrapContext instanceof ContextWrapper) { unwrapContext = ((ContextWrapper) unwrapContext).getBaseContext(); if (unwrapContext instanceof Application) { return (Application) unwrapContext; } } throw new IllegalStateException( "Could not find an Application in the given context: " + context); } private Contexts() {} }
406
10,225
package org.jboss.resteasy.reactive.common.providers.serialisers; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.charset.StandardCharsets; import javax.ws.rs.core.MediaType; public class MessageReaderUtil { public static final String UTF8_CHARSET = StandardCharsets.UTF_8.name(); public static String charsetFromMediaType(MediaType mediaType) { if (mediaType == null) { return UTF8_CHARSET; } String charset = mediaType.getParameters().get(MediaType.CHARSET_PARAMETER); if (charset != null) { return charset; } return UTF8_CHARSET; } public static byte[] readBytes(InputStream entityStream) throws IOException { ByteArrayOutputStream out = new ByteArrayOutputStream(); byte[] buf = new byte[1024]; //TODO: fix, needs a pure vert.x async read model int r; while ((r = entityStream.read(buf)) > 0) { out.write(buf, 0, r); } return out.toByteArray(); } public static String readString(InputStream entityStream, MediaType mediaType) throws IOException { return new String(readBytes(entityStream), charsetFromMediaType(mediaType)); } }
492
310
<gh_stars>100-1000 { "name": "import.io", "description": "A service for scraping data from websites.", "url": "https://www.import.io/" }
50
2,462
package com.sanshengshui.persistence; import com.sanshengshui.persistence.entity.ToData; import java.util.*; public abstract class DaoUtil { private DaoUtil() { } public static <T> List<T> convertDataList(Collection<? extends ToData<T>> toDataList) { List<T> list = Collections.emptyList(); if (toDataList != null && !toDataList.isEmpty()) { list = new ArrayList<>(); for (ToData<T> object : toDataList) { if (object != null) { list.add(object.toData()); } } } return list; } public static <T> T getData(ToData<T> data) { T object = null; if (data != null) { object = data.toData(); } return object; } public static Long getId(Long idBased) { Long id = null; if (idBased != null) { id = idBased; } return id; } public static List<Long> toUUIDs(List<? extends Long> idBasedIds) { List<Long> ids = new ArrayList<>(); for (Long idBased : idBasedIds) { ids.add(getId(idBased)); } return ids; } }
591
679
<gh_stars>100-1000 /************************************************************** * * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. * *************************************************************/ package org.openoffice.xmerge.converter.xml; import org.w3c.dom.Node; import org.w3c.dom.NamedNodeMap; /** * An object of class <code>Style</code> represents a <i>style</i> * in an OpenOffice document. In practice subclasses of this * <code>Style</code>, such as <code>TextStyle</code>, * <code>ParaStyle</code> are used. * * @author <NAME> * @see <a href="TextStyle.html">TextStyle</a>, * <a href="ParaStyle.html">ParaStyle</a> */ public class Style { /** Name of the <code>Style</code>. */ protected String name = null; /** Family of the <code>Style</code>. */ protected String family = null; /** Parent of the <code>Style</code>. */ protected String parent = null; /** * A reference to the <code>StyleCatalog</code> to be used for * looking up ancestor <code>Style</code> objects. */ protected StyleCatalog sc; /** * Constructor for use when going from DOM to client device format. * * @param node A <i>style:style</i> or <i>style:default-style</i> * <code>Node</code> from the document being parsed. * No checking of <code>Node</code> is done, so if it * is not of the proper type the results will be * unpredictable. * @param sc The <code>StyleCatalog</code>, which is used for * looking up ancestor <code>Style</code> objects. */ public Style(Node node, StyleCatalog sc) { this.sc = sc; // Run through the attributes of this node, saving // the ones we're interested in. if (node.getNodeName().equals("style:default-style")) name = "DEFAULT_STYLE"; NamedNodeMap attrNodes = node.getAttributes(); if (attrNodes != null) { int len = attrNodes.getLength(); for (int i = 0; i < len; i++) { Node attr = attrNodes.item(i); if (attr.getNodeName().equals("style:family")) family = attr.getNodeValue(); else if (attr.getNodeName().equals("style:name")) { name = attr.getNodeValue(); } else if (attr.getNodeName().equals("style:parent-style-name")) parent = attr.getNodeValue(); } } } /** * Constructor for use when going from client device format to DOM. * * @param name Name of the <code>Style</code>. Can be null. * @param family Family of the <code>Style</code> - usually * <i>paragraph</i>, <i>text</i>, etc. Can be null. * @param parent Name of the parent <code>Style</code>, or null if none. * @param sc The <code>StyleCatalog</code>, which is used for * looking up ancestor <code>Style</code> objects. */ public Style(String name, String family, String parent, StyleCatalog sc) { this.sc = sc; this.name = name; this.family = family; this.parent = parent; } /** * Set the <code>StyleCatalog</code> to be used when looking up the * <code>Style</code> parent. * * @param sc The <code>StyleCatalog</code>, which is used for * looking up ancestor <code>Style</code> objects. */ public void setCatalog(StyleCatalog sc) { this.sc = sc; } /** * Returns the name of this <code>Style</code>. * * @return The name of this <code>Style</code>. */ public String getName() { return name; } /** * Sets the name of this <code>Style</code>. * * @param newName The new name of this <code>Style</code>. */ public void setName(String newName) { name = newName; } /** * Return the family of this <code>Style</code>. * * @return The family of this <code>Style</code>. */ public String getFamily() { return family; } /** * Return the name of the parent of this <code>Style</code>. * * @return The parent of this <code>Style</code>. */ public String getParent() { return parent; } /** * Return a <code>Style</code> object corresponding to this one, but with * all of the inherited information from parent <code>Style</code> * objects filled in. The object returned will be a new object, not a * reference to this object, even if it does not need any information * added. * * @return A resolved <code>Style</code> object in which to look up * ancestors. */ public Style getResolved() { return new Style(name, family, parent, sc); } /** * Write a <code>Node</code> in <code>parentDoc</code> * representing this <code>Style</code>. Note that the * <code>Node</code> is returned unconnected. * * @param parentDoc Document to which new <code>Node</code> will * belong. * @param name Name to use for new <code>Node</code>. */ public Node createNode(org.w3c.dom.Document parentDoc, String name) { // DJP: write this! Should call writeAttributes() return null; } /** * Write this <code>Style</code> object's attributes to the given * <code>Node</code>. This may involve writing child * <code>Node</code> objects as well. This is similar to the * <code>writeNode</code> method, but the <code>Node</code> * already exists, and this does <b>not</b> write the name, * family, and parent attributes, which are assumed to already * exist in the <code>Node</code>. * * @param node The <code>Node</code> to add style attributes. */ public void writeAttributes(Node node) { } /** * Return true if <code>Style</code> is a subset of this one. Note * that this will return true even if <code>Style</code> is less * specific than this <code>Style</code>, so long as it does not * contradict this <code>Style</code> in any way. * * This always returns true since only subclasses of * <code>Style</code> contain any actual <code>Style</code> * information. * * @param style The <code>Style</code> to check * * @return true if the <code>Style</code> is a subset, false otherwise. */ public boolean isSubset(Style style) { return true; } }
3,019
435
<reponame>amaajemyfren/data<gh_stars>100-1000 { "alias": "video/1480/python-web-sprints", "category": "PyCon DE 2012", "copyright_text": "", "description": "", "duration": null, "id": 1480, "language": "deu", "quality_notes": "", "recorded": "2012-11-01", "slug": "python-web-sprints", "speakers": [ "<NAME>", "<NAME>" ], "summary": "Seit April 2011 finden regelm\u00e4\u00dfig dreit\u00e4gige Sprints zu verschiedenen\nThemen rund um Python-Web Anwendungen statt. Dabei geht es meist nicht\num die schnelle Implementierung spezifischer Features sondern um\ngrundlegende, von den Beteiligten mitgetragene Konzepte. Meist wird auch\nnicht an einzelnen Komponenten, sondern an einer Toolchain zur L\u00f6sung\nkonkreter Probleme gearbeitet. Schlie\u00dflich werden die Ergebnisse in\neiner Anwenderdokumentation festgehalten.\n\nDie Initiatoren des Sprints, <NAME> und <NAME>, stellen\ndas Konzept der Sprint-Reihe vor und gehen auch die Ergebnisse der\nbisherigen Sprints ein. Die Themen der bisherigen Sprints waren:\nPyPI-Mirroring und Performance-Optimierungen, Sicherheit und Datenschutz\nbei Web-Anwendungen sowie Betrieb und Deployment von\nPython-Webanwendungen.\n", "tags": [ "sprint", "web-anwendungen" ], "thumbnail_url": "https://i4.ytimg.com/vi/sX5zuqomc14/hqdefault.jpg", "title": "Python-Web-Sprints", "videos": [ { "type": "mp4", "url": "http://s3.us.archive.org/nextdayvideo/pyconde/pyconde2012/Vortrag_PythonWebSprints.mp4?Signature=kmgV9ltUeZT%2B4SIor4Lqna770PQ%3D&Expires=1352081375&AWSAccessKeyId=<KEY>" }, { "length": 0, "type": "youtube", "url": "https://www.youtube.com/watch?v=sX5zuqomc14" } ] }
742
2,329
package fields; public class Ja extends Ia { public int getField() { return field; } public void setField(int newvalue) { field = newvalue; } }
56
337
<gh_stars>100-1000 //statement Character i = 10;
17
4,054
// Copyright Yahoo. Licensed under the terms of the Apache 2.0 license. See LICENSE in the project root. package com.yahoo.vespa.hosted.controller.api.integration.noderepository; import com.fasterxml.jackson.annotation.JsonCreator; import com.fasterxml.jackson.annotation.JsonIgnoreProperties; import com.fasterxml.jackson.annotation.JsonInclude; import com.fasterxml.jackson.annotation.JsonProperty; /** * Patchable data under Application * * @author bratseth */ @JsonIgnoreProperties(ignoreUnknown = true) @JsonInclude(JsonInclude.Include.NON_NULL) public class ApplicationPatch { @JsonProperty private final Double currentReadShare; @JsonProperty private final Double maxReadShare; @JsonCreator public ApplicationPatch(@JsonProperty("currentReadShare") Double currentReadShare, @JsonProperty("maxReadShare") Double maxReadShare) { this.currentReadShare = currentReadShare; this.maxReadShare = maxReadShare; } public Double getCurrentReadShare() { return currentReadShare; } public Double getMaxReadShare() { return maxReadShare; } }
381
852
import FWCore.ParameterSet.Config as cms # # Strip calib # from CalibTracker.Configuration.SiStripCabling.SiStripCabling_Frontier_cff import * from CalibTracker.Configuration.SiStripGain.SiStripGain_Frontier_cff import * from CalibTracker.Configuration.SiStripLorentzAngle.SiStripLorentzAngle_Frontier_cff import * from CalibTracker.Configuration.SiStripNoise.SiStripNoise_Frontier_cff import * from CalibTracker.Configuration.SiStripPedestals.SiStripPedestals_Fake_cff import * from CalibTracker.Configuration.SiStripQuality.SiStripQuality_Fake_cff import *
189
421
// <Snippet3> #using <system.dll> using namespace System; public ref class HelloServiceClass: public MarshalByRefObject { private: static int n_instance; public: HelloServiceClass() { n_instance++; Console::WriteLine( "{0} has been created. Instance # = {1}", this->GetType()->Name, n_instance ); } ~HelloServiceClass() { Console::WriteLine( "Destroyed instance {0} of HelloServiceClass.", n_instance ); n_instance--; } String^ HelloMethod( String^ name ) { // Reports that the method was called. Console::WriteLine(); Console::WriteLine( "Called HelloMethod on instance {0} with the '{1}' parameter.", n_instance, name ); // Calculates and returns the result to the client. return String::Format( "Hi there {0}", name ); } }; // </Snippet3>
342
1,455
<filename>src/main/java/org/springframework/data/redis/support/atomic/RedisAtomicDouble.java /* * Copyright 2013-2021 the original author or authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * https://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.springframework.data.redis.support.atomic; import java.io.Serializable; import java.util.Date; import java.util.concurrent.TimeUnit; import java.util.function.DoubleBinaryOperator; import java.util.function.DoubleUnaryOperator; import org.springframework.dao.DataRetrievalFailureException; import org.springframework.data.redis.connection.DataType; import org.springframework.data.redis.connection.RedisConnectionFactory; import org.springframework.data.redis.core.BoundKeyOperations; import org.springframework.data.redis.core.RedisOperations; import org.springframework.data.redis.core.RedisTemplate; import org.springframework.data.redis.core.ValueOperations; import org.springframework.data.redis.serializer.GenericToStringSerializer; import org.springframework.data.redis.serializer.RedisSerializer; import org.springframework.lang.Nullable; import org.springframework.util.Assert; /** * Atomic double backed by Redis. Uses Redis atomic increment/decrement and watch/multi/exec operations for CAS * operations. * * @author <NAME> * @author <NAME> * @author <NAME> * @author <NAME> * @author <NAME> * @author <NAME> */ public class RedisAtomicDouble extends Number implements Serializable, BoundKeyOperations<String> { private static final long serialVersionUID = 1L; private volatile String key; private final ValueOperations<String, Double> operations; private final RedisOperations<String, Double> generalOps; /** * Constructs a new {@link RedisAtomicDouble} instance. Uses the value existing in Redis or {@code 0} if none is * found. * * @param redisCounter Redis key of this counter. * @param factory connection factory. */ public RedisAtomicDouble(String redisCounter, RedisConnectionFactory factory) { this(redisCounter, factory, null); } /** * Constructs a new {@link RedisAtomicDouble} instance with a {@code initialValue} that overwrites the existing value. * * @param redisCounter Redis key of this counter. * @param factory connection factory. * @param initialValue initial value to set. */ public RedisAtomicDouble(String redisCounter, RedisConnectionFactory factory, double initialValue) { this(redisCounter, factory, Double.valueOf(initialValue)); } private RedisAtomicDouble(String redisCounter, RedisConnectionFactory factory, @Nullable Double initialValue) { Assert.hasText(redisCounter, "a valid counter name is required"); Assert.notNull(factory, "a valid factory is required"); RedisTemplate<String, Double> redisTemplate = new RedisTemplate<>(); redisTemplate.setKeySerializer(RedisSerializer.string()); redisTemplate.setValueSerializer(new GenericToStringSerializer<>(Double.class)); redisTemplate.setExposeConnection(true); redisTemplate.setConnectionFactory(factory); redisTemplate.afterPropertiesSet(); this.key = redisCounter; this.generalOps = redisTemplate; this.operations = generalOps.opsForValue(); if (initialValue == null) { initializeIfAbsent(); } else { set(initialValue); } } /** * Constructs a new {@link RedisAtomicDouble} instance. Uses the value existing in Redis or 0 if none is found. * * @param redisCounter Redis key of this counter. * @param template the template. * @see #RedisAtomicDouble(String, RedisConnectionFactory, double) */ public RedisAtomicDouble(String redisCounter, RedisOperations<String, Double> template) { this(redisCounter, template, null); } /** * Constructs a new {@link RedisAtomicDouble} instance with a {@code initialValue} that overwrites the existing value * at {@code redisCounter}. * <p> * Note: You need to configure the given {@code template} with appropriate {@link RedisSerializer} for the key and * value. * <p> * As an alternative one could use the {@link #RedisAtomicDouble(String, RedisConnectionFactory, Double)} constructor * which uses appropriate default serializers. * * @param redisCounter Redis key of this counter. * @param template the template * @param initialValue initial value to set if the Redis key is absent. */ public RedisAtomicDouble(String redisCounter, RedisOperations<String, Double> template, double initialValue) { this(redisCounter, template, Double.valueOf(initialValue)); } private RedisAtomicDouble(String redisCounter, RedisOperations<String, Double> template, @Nullable Double initialValue) { Assert.hasText(redisCounter, "a valid counter name is required"); Assert.notNull(template, "a valid template is required"); Assert.notNull(template.getKeySerializer(), "a valid key serializer in template is required"); Assert.notNull(template.getValueSerializer(), "a valid value serializer in template is required"); this.key = redisCounter; this.generalOps = template; this.operations = generalOps.opsForValue(); if (initialValue == null) { initializeIfAbsent(); } else { set(initialValue); } } private void initializeIfAbsent() { operations.setIfAbsent(key, (double) 0); } /** * Get the current value. * * @return the current value. */ public double get() { Double value = operations.get(key); if (value != null) { return value; } throw new DataRetrievalFailureException(String.format("The key '%s' seems to no longer exist.", key)); } /** * Set to the given value. * * @param newValue the new value. */ public void set(double newValue) { operations.set(key, newValue); } /** * Set to the given value and return the old value. * * @param newValue the new value. * @return the previous value. */ public double getAndSet(double newValue) { Double value = operations.getAndSet(key, newValue); return value != null ? value : 0; } /** * Atomically set the value to the given updated value if the current value {@code ==} the expected value. * * @param expect the expected value. * @param update the new value. * @return {@literal true} if successful. {@literal false} indicates that the actual value was not equal to the * expected value. */ public boolean compareAndSet(double expect, double update) { return generalOps.execute(new CompareAndSet<>(this::get, this::set, key, expect, update)); } /** * Atomically increment by one the current value. * * @return the previous value. */ public double getAndIncrement() { return incrementAndGet() - 1.0; } /** * Atomically decrement by one the current value. * * @return the previous value. */ public double getAndDecrement() { return decrementAndGet() + 1.0; } /** * Atomically add the given value to current value. * * @param delta the value to add. * @return the previous value. */ public double getAndAdd(double delta) { return addAndGet(delta) - delta; } /** * Atomically update the current value using the given {@link DoubleUnaryOperator update function}. * * @param updateFunction the function which calculates the value to set. Should be a pure function (no side effects), * because it will be applied several times if update attempts fail due to concurrent calls. Must not be * {@literal null}. * @return the previous value. * @since 2.2 */ public double getAndUpdate(DoubleUnaryOperator updateFunction) { Assert.notNull(updateFunction, "Update function must not be null!"); double previousValue, newValue; do { previousValue = get(); newValue = updateFunction.applyAsDouble(previousValue); } while (!compareAndSet(previousValue, newValue)); return previousValue; } /** * Atomically update the current value using the given {@link DoubleBinaryOperator accumulator function}. The new * value is calculated by applying the accumulator function to the current value and the given {@code updateValue}. * * @param updateValue the value which will be passed into the accumulator function. * @param accumulatorFunction the function which calculates the value to set. Should be a pure function (no side * effects), because it will be applied several times if update attempts fail due to concurrent calls. Must * not be {@literal null}. * @return the previous value. * @since 2.2 */ public double getAndAccumulate(double updateValue, DoubleBinaryOperator accumulatorFunction) { Assert.notNull(accumulatorFunction, "Accumulator function must not be null!"); double previousValue, newValue; do { previousValue = get(); newValue = accumulatorFunction.applyAsDouble(previousValue, updateValue); } while (!compareAndSet(previousValue, newValue)); return previousValue; } /** * Atomically increment by one the current value. * * @return the updated value. */ public double incrementAndGet() { return operations.increment(key, 1.0); } /** * Atomically decrement by one the current value. * * @return the updated value. */ public double decrementAndGet() { return operations.increment(key, -1.0); } /** * Atomically add the given value to current value. * * @param delta the value to add. * @return the updated value. */ public double addAndGet(double delta) { return operations.increment(key, delta); } /** * Atomically update the current value using the given {@link DoubleUnaryOperator update function}. * * @param updateFunction the function which calculates the value to set. Should be a pure function (no side effects), * because it will be applied several times if update attempts fail due to concurrent calls. Must not be * {@literal null}. * @return the updated value. * @since 2.2 */ public double updateAndGet(DoubleUnaryOperator updateFunction) { Assert.notNull(updateFunction, "Update function must not be null!"); double previousValue, newValue; do { previousValue = get(); newValue = updateFunction.applyAsDouble(previousValue); } while (!compareAndSet(previousValue, newValue)); return newValue; } /** * Atomically update the current value using the given {@link DoubleBinaryOperator accumulator function}. The new * value is calculated by applying the accumulator function to the current value and the given {@code updateValue}. * * @param updateValue the value which will be passed into the accumulator function. * @param accumulatorFunction the function which calculates the value to set. Should be a pure function (no side * effects), because it will be applied several times if update attempts fail due to concurrent calls. Must * not be {@literal null}. * @return the updated value. * @since 2.2 */ public double accumulateAndGet(double updateValue, DoubleBinaryOperator accumulatorFunction) { Assert.notNull(accumulatorFunction, "Accumulator function must not be null!"); double previousValue, newValue; do { previousValue = get(); newValue = accumulatorFunction.applyAsDouble(previousValue, updateValue); } while (!compareAndSet(previousValue, newValue)); return newValue; } /** * @return the String representation of the current value. */ @Override public String toString() { return Double.toString(get()); } /* * (non-Javadoc) * @see org.springframework.data.redis.core.BoundKeyOperations#getKey() */ @Override public String getKey() { return key; } /* * (non-Javadoc) * @see org.springframework.data.redis.core.BoundKeyOperations#getType() */ @Override public DataType getType() { return DataType.STRING; } /* * (non-Javadoc) * @see org.springframework.data.redis.core.BoundKeyOperations#getExpire() */ @Override public Long getExpire() { return generalOps.getExpire(key); } /* * (non-Javadoc) * @see org.springframework.data.redis.core.BoundKeyOperations#expire(long, java.util.concurrent.TimeUnit) */ @Override public Boolean expire(long timeout, TimeUnit unit) { return generalOps.expire(key, timeout, unit); } /* * (non-Javadoc) * @see org.springframework.data.redis.core.BoundKeyOperations#expireAt(java.util.Date) */ @Override public Boolean expireAt(Date date) { return generalOps.expireAt(key, date); } /* * (non-Javadoc) * @see org.springframework.data.redis.core.BoundKeyOperations#persist() */ @Override public Boolean persist() { return generalOps.persist(key); } /* * (non-Javadoc) * @see org.springframework.data.redis.core.BoundKeyOperations#rename(java.lang.Object) */ @Override public void rename(String newKey) { generalOps.rename(key, newKey); key = newKey; } /* * (non-Javadoc) * @see java.lang.Number#intValue() */ @Override public int intValue() { return (int) get(); } /* * (non-Javadoc) * @see java.lang.Number#longValue() */ @Override public long longValue() { return (long) get(); } /* * (non-Javadoc) * @see java.lang.Number#floatValue() */ @Override public float floatValue() { return (float) get(); } /* * (non-Javadoc) * @see java.lang.Number#doubleValue() */ @Override public double doubleValue() { return get(); } }
4,310
930
package com.foxinmy.weixin4j.mp.component; import java.io.Serializable; import java.util.Date; import javax.xml.bind.annotation.XmlAccessType; import javax.xml.bind.annotation.XmlAccessorType; import javax.xml.bind.annotation.XmlElement; import javax.xml.bind.annotation.XmlRootElement; import javax.xml.bind.annotation.XmlTransient; /** * 组件消息 * * @className ComponentMessage * @author jinyu(<EMAIL>) * @date 2016年7月5日 * @since JDK 1.6 */ @XmlRootElement(name = "xml") @XmlAccessorType(XmlAccessType.FIELD) public class ComponentMessage implements Serializable { private static final long serialVersionUID = -7243616276403632118L; /** * 第三方平台appid */ @XmlElement(name = "AppId") private String appId; /** * 事件类型 */ @XmlElement(name = "InfoType") private String eventType; /** * 时间戳 */ @XmlElement(name = "CreateTime") private long createTime; /** * Ticket内容 */ @XmlElement(name = "ComponentVerifyTicket") private String verifyTicket; /** * 授权方的Appid */ @XmlElement(name = "AuthorizerAppid") private String authAppId; /** * 授权码,可用于换取公众号的接口调用凭据 */ @XmlElement(name = "AuthorizationCode") private String authCode; /** * 授权码过期时间 */ @XmlElement(name = "AuthorizationCodeExpiredTime") private long authCodeExpiredTime; public String getAppId() { return appId; } public String getEventType() { return eventType; } @XmlTransient public ComponentEventType getFormatEventType() { return ComponentEventType.valueOf(eventType); } public long getCreateTime() { return createTime; } @XmlTransient public Date getFormatCreateTime() { return createTime > 0l ? new Date(createTime * 1000l) : null; } public String getVerifyTicket() { return verifyTicket; } public String getAuthAppId() { return authAppId; } public String getAuthCode() { return authCode; } public long getAuthCodeExpiredTime() { return authCodeExpiredTime; } @XmlTransient public Date getFormatAuthCodeExpiredTime() { return authCodeExpiredTime > 0l ? new Date(authCodeExpiredTime * 1000l) : null; } @Override public String toString() { return "ComponentMessage [appId=" + appId + ", eventType=" + eventType + ", createTime=" + createTime + ", verifyTicket=" + verifyTicket + ", authAppId=" + authAppId + ", authCode=" + authCode + ", authCodeExpiredTime=" + authCodeExpiredTime + "]"; } }
943
7,137
package io.onedev.server.util.concurrent; import java.util.concurrent.FutureTask; public class PrioritizedFutureTask<T> extends FutureTask<T> implements PriorityAware, Comparable<PriorityAware> { private final int priority; public PrioritizedFutureTask(PrioritizedCallable<T> callable) { super(callable); this.priority = callable.getPriority(); } public PrioritizedFutureTask(PrioritizedRunnable runnable, T value) { super(runnable, value); this.priority = runnable.getPriority(); } @Override public int getPriority() { return priority; } @Override public int compareTo(PriorityAware o) { return priority - o.getPriority(); } }
229
1,626
<filename>test/sample.cpp // // Copyright (c) 2016, Scientific Toolworks, Inc. // // This software is licensed under the MIT License. The LICENSE.md file // describes the conditions under which this software may be distributed. // // Author: <NAME> // #include "Test.h" #include <QMainWindow> using namespace Test; using namespace QTest; class TestSample : public QObject { Q_OBJECT private slots: void initTestCase(); void sample(); void cleanupTestCase(); private: int inputDelay = 99; int closeDelay = 9999; ScratchRepository *mRepo = nullptr; QMainWindow *mWindow = nullptr; }; void TestSample::initTestCase() { mRepo = new ScratchRepository; mWindow = new QMainWindow; //UiComponent *ui = new UiComponent(mRepo->repo(), mWindow); //ui->show(); //QVERIFY(qWaitForWindowActive(ui)); } void TestSample::sample() { //keyClick(ui, 'Qt::Key_Return', Qt::NoModifier, inputDelay); } void TestSample::cleanupTestCase() { qWait(closeDelay); mWindow->close(); delete mRepo; } TEST_MAIN(TestSample) #include "sample.moc"
381
14,668
// Copyright 2020 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #ifndef CHROME_BROWSER_CHROMEOS_FILEAPI_FILE_CHANGE_SERVICE_OBSERVER_H_ #define CHROME_BROWSER_CHROMEOS_FILEAPI_FILE_CHANGE_SERVICE_OBSERVER_H_ #include "base/observer_list_types.h" namespace storage { class FileSystemURL; } // namespace storage namespace chromeos { // An interface for an observer which receives `FileChangeService` events. class FileChangeServiceObserver : public base::CheckedObserver { public: // Invoked when a file identified by `url` has been modified. Note that this // will not get called on file creation or deletion. virtual void OnFileModified(const storage::FileSystemURL& url) {} // Invoked when a file has been copied from `src` to `dst`. virtual void OnFileCopied(const storage::FileSystemURL& src, const storage::FileSystemURL& dst) {} // Invoked when a file has been moved from `src` to `dst`. virtual void OnFileMoved(const storage::FileSystemURL& src, const storage::FileSystemURL& dst) {} }; } // namespace chromeos #endif // CHROME_BROWSER_CHROMEOS_FILEAPI_FILE_CHANGE_SERVICE_OBSERVER_H_
433
2,342
/*********************************************************************** Vczh Library++ 3.0 Developer: <NAME>(vczh) GacUI::Control Styles::Windows7 Styles Clases: ***********************************************************************/ #ifndef VCZH_PRESENTATION_CONTROLS_WIN7STYLES_GUIWIN7STYLESCOMMON #define VCZH_PRESENTATION_CONTROLS_WIN7STYLES_GUIWIN7STYLESCOMMON #include "../GuiCommonStyles.h" namespace vl { namespace presentation { namespace win7 { /*********************************************************************** Button Configuration ***********************************************************************/ struct Win7ButtonColors { Color borderColor; Color backgroundColor; Color g1; Color g2; Color g3; Color g4; Color textColor; Color bulletLight; Color bulletDark; bool operator==(const Win7ButtonColors& colors) { return borderColor == colors.borderColor && backgroundColor == colors.backgroundColor && g1 == colors.g1 && g2 == colors.g2 && g3 == colors.g3 && g4 == colors.g4 && textColor == colors.textColor && bulletLight == colors.bulletLight && bulletDark == colors.bulletDark; } bool operator!=(const Win7ButtonColors& colors) { return !(*this==colors); } void SetAlphaWithoutText(unsigned char a); static Win7ButtonColors Blend(const Win7ButtonColors& c1, const Win7ButtonColors& c2, vint ratio, vint total); static Win7ButtonColors ButtonNormal(); static Win7ButtonColors ButtonActive(); static Win7ButtonColors ButtonPressed(); static Win7ButtonColors ButtonDisabled(); static Win7ButtonColors ItemNormal(); static Win7ButtonColors ItemActive(); static Win7ButtonColors ItemSelected(); static Win7ButtonColors ItemDisabled(); static Win7ButtonColors CheckedNormal(bool selected); static Win7ButtonColors CheckedActive(bool selected); static Win7ButtonColors CheckedPressed(bool selected); static Win7ButtonColors CheckedDisabled(bool selected); static Win7ButtonColors ToolstripButtonNormal(); static Win7ButtonColors ToolstripButtonActive(); static Win7ButtonColors ToolstripButtonPressed(); static Win7ButtonColors ToolstripButtonSelected(); static Win7ButtonColors ToolstripButtonDisabled(); static Win7ButtonColors MenuBarButtonNormal(); static Win7ButtonColors MenuBarButtonActive(); static Win7ButtonColors MenuBarButtonPressed(); static Win7ButtonColors MenuBarButtonDisabled(); static Win7ButtonColors MenuItemButtonNormal(); static Win7ButtonColors MenuItemButtonNormalActive(); static Win7ButtonColors MenuItemButtonSelected(); static Win7ButtonColors MenuItemButtonSelectedActive(); static Win7ButtonColors MenuItemButtonDisabled(); static Win7ButtonColors MenuItemButtonDisabledActive(); static Win7ButtonColors TabPageHeaderNormal(); static Win7ButtonColors TabPageHeaderActive(); static Win7ButtonColors TabPageHeaderSelected(); }; struct Win7ButtonElements { elements::GuiSolidBorderElement* rectBorderElement; elements::GuiRoundBorderElement* roundBorderElement; elements::GuiSolidBackgroundElement* backgroundElement; elements::GuiGradientBackgroundElement* topGradientElement; elements::GuiGradientBackgroundElement* bottomGradientElement; elements::GuiSolidLabelElement* textElement; compositions::GuiBoundsComposition* textComposition; compositions::GuiBoundsComposition* mainComposition; compositions::GuiBoundsComposition* backgroundComposition; compositions::GuiTableComposition* gradientComposition; static Win7ButtonElements Create(bool verticalGradient, bool roundBorder, Alignment horizontal=Alignment::Center, Alignment vertical=Alignment::Center); void Apply(const Win7ButtonColors& colors); }; struct Win7CheckedButtonElements { elements::GuiSolidBorderElement* borderElement; elements::GuiSolidBackgroundElement* backgroundElement; elements::GuiGradientBackgroundElement* outerGradientElement; elements::GuiGradientBackgroundElement* innerGradientElement; elements::GuiSolidLabelElement* textElement; elements::GuiSolidLabelElement* bulletCheckElement; elements::GuiSolidBackgroundElement* bulletRadioElement; compositions::GuiBoundsComposition* textComposition; compositions::GuiBoundsComposition* mainComposition; static Win7CheckedButtonElements Create(elements::ElementShape shape, bool backgroundVisible); void Apply(const Win7ButtonColors& colors); }; struct Win7MenuItemButtonElements { elements::GuiRoundBorderElement* borderElement; elements::GuiSolidBackgroundElement* backgroundElement; elements::GuiGradientBackgroundElement* gradientElement; elements::Gui3DSplitterElement* splitterElement; compositions::GuiCellComposition* splitterComposition; elements::GuiImageFrameElement* imageElement; elements::GuiSolidLabelElement* textElement; compositions::GuiSharedSizeItemComposition* textComposition; elements::GuiSolidLabelElement* shortcutElement; compositions::GuiSharedSizeItemComposition* shortcutComposition; elements::GuiPolygonElement* subMenuArrowElement; compositions::GuiGraphicsComposition* subMenuArrowComposition; compositions::GuiBoundsComposition* mainComposition; static Win7MenuItemButtonElements Create(); void Apply(const Win7ButtonColors& colors); void SetActive(bool value); void SetSubMenuExisting(bool value); }; struct Win7TextBoxColors { Color borderColor; Color backgroundColor; bool operator==(const Win7TextBoxColors& colors) { return borderColor == colors.borderColor && backgroundColor == colors.backgroundColor; } bool operator!=(const Win7TextBoxColors& colors) { return !(*this==colors); } static Win7TextBoxColors Blend(const Win7TextBoxColors& c1, const Win7TextBoxColors& c2, vint ratio, vint total); static Win7TextBoxColors Normal(); static Win7TextBoxColors Active(); static Win7TextBoxColors Focused(); static Win7TextBoxColors Disabled(); }; /*********************************************************************** Helper Functions ***********************************************************************/ extern Color Win7GetSystemWindowColor(); extern Color Win7GetSystemTabContentColor(); extern Color Win7GetSystemBorderColor(); extern Color Win7GetSystemBorderSinkColor(); extern Color Win7GetSystemBorderRaiseColor(); extern Color Win7GetSystemTextColor(bool enabled); extern void Win7SetFont(elements::GuiSolidLabelElement* element, compositions::GuiBoundsComposition* composition, const FontProperties& fontProperties); extern void Win7CreateSolidLabelElement(elements::GuiSolidLabelElement*& element, compositions::GuiBoundsComposition*& composition, Alignment horizontal, Alignment vertical); extern void Win7CreateSolidLabelElement(elements::GuiSolidLabelElement*& element, compositions::GuiSharedSizeItemComposition*& composition, const WString& group, Alignment horizontal, Alignment vertical); extern elements::text::ColorEntry Win7GetTextBoxTextColor(); } } } #endif
2,994
409
<reponame>jtavara23/cassandra-workshop-series<filename>week4-AppDev-api/samples-codes/sample-kafka/kafka-dse-webui/src/main/java/com/datastax/demo/springdata/TickSpringDataRepository.java package com.datastax.demo.springdata; import com.datastax.demo.conf.DseConstants; import com.datastax.demo.springdata.dto.TickData; import com.datastax.demo.springdata.dto.TickDataPrimaryKey; import org.springframework.data.cassandra.repository.Query; import org.springframework.data.repository.reactive.ReactiveCrudRepository; import reactor.core.publisher.Flux; /** Retrieve all Tick from Table. */ public interface TickSpringDataRepository extends ReactiveCrudRepository<TickData, TickDataPrimaryKey> { @Query("SELECT * FROM " + DseConstants.STOCKS_TICKS + " WHERE symbol = ?0 LIMIT 100") Flux<TickData> findBySymbol(String symbol); @Query( "SELECT * FROM " + DseConstants.STOCKS_TICKS + " WHERE symbol in ('BAC', 'DVMT','DIS','IBM','WMT') LIMIT 500") Flux<TickData> findAllLastSymbols(); }
392
892
<filename>advisories/unreviewed/2022/05/GHSA-3j75-qm7x-j2gc/GHSA-3j75-qm7x-j2gc.json { "schema_version": "1.2.0", "id": "GHSA-3j75-qm7x-j2gc", "modified": "2022-05-01T23:47:37Z", "published": "2022-05-01T23:47:37Z", "aliases": [ "CVE-2008-2200" ], "details": "Multiple cross-site scripting (XSS) vulnerabilities in Maian Weblog 4.0 allow remote attackers to inject arbitrary web script or HTML via the (1) keywords parameter to admin/index.php in a blogs search action, the (2) msg_charset and (3) msg_header9 parameters to admin/inc/header.php, and the (4) keywords parameter to index.php in a search action.", "severity": [ ], "affected": [ ], "references": [ { "type": "ADVISORY", "url": "https://nvd.nist.gov/vuln/detail/CVE-2008-2200" }, { "type": "WEB", "url": "https://exchange.xforce.ibmcloud.com/vulnerabilities/42207" }, { "type": "WEB", "url": "http://secunia.com/advisories/30060" }, { "type": "WEB", "url": "http://securityreason.com/securityalert/3880" }, { "type": "WEB", "url": "http://www.securityfocus.com/archive/1/491588/100/0/threaded" }, { "type": "WEB", "url": "http://www.securityfocus.com/bid/29032" } ], "database_specific": { "cwe_ids": [ "CWE-79" ], "severity": "MODERATE", "github_reviewed": false } }
644
334
from calvin.utilities.utils import enum # Only TEMPORARY, TERMINATE and EXHAUST are used on highest level in actor and port manager # TEMPORARY is used during migration # TERMINATE is used for dereplication without preserving tokens # EXHAUST is used for dereplication while preserving tokens # The different exhaust disconnects are based on a higher level exhaust value # EXHAUST and EXHAUST_PEER are used at actor port, endpoint and connection level # EXHAUST_INPORT, EXHAUST_OUTPORT, EXHAUST_PEER_SEND, EXHAUST_PEER_RECV are used at queue level DISCONNECT = enum('TEMPORARY', 'TERMINATE', 'EXHAUST', 'EXHAUST_PEER', 'EXHAUST_INPORT', 'EXHAUST_OUTPORT', 'EXHAUST_PEER_RECV', 'EXHAUST_PEER_SEND')
239
416
/* File: AVAudioChannelLayout.h Framework: AVFoundation Copyright (c) 2014-2015 Apple Inc. All Rights Reserved. */ #import <AVFAudio/AVAudioTypes.h> #import <CoreAudio/CoreAudioTypes.h> NS_ASSUME_NONNULL_BEGIN /*! @class AVAudioChannelLayout @abstract A description of the roles of a set of audio channels. @discussion This object is a thin wrapper for the AudioChannelLayout structure, described in <CoreAudio/CoreAudioTypes.h>. */ OS_EXPORT API_AVAILABLE(macos(10.10), ios(8.0), watchos(2.0), tvos(9.0)) @interface AVAudioChannelLayout : NSObject <NSSecureCoding> { @private AudioChannelLayoutTag _layoutTag; AudioChannelLayout * _layout; void *_reserved; } - (instancetype)init NS_UNAVAILABLE; /*! @method initWithLayoutTag: @abstract Initialize from a layout tag. @param layoutTag The tag. @discussion Returns nil if the tag is either kAudioChannelLayoutTag_UseChannelDescriptions or kAudioChannelLayoutTag_UseChannelBitmap. */ - (nullable instancetype)initWithLayoutTag:(AudioChannelLayoutTag)layoutTag; /*! @method initWithLayout: @abstract Initialize from an AudioChannelLayout. @param layout The AudioChannelLayout. @discussion If the provided layout's tag is kAudioChannelLayoutTag_UseChannelDescriptions, this initializer attempts to convert it to a more specific tag. */ - (instancetype)initWithLayout:(const AudioChannelLayout *)layout NS_DESIGNATED_INITIALIZER; /*! @method isEqual: @abstract Determine whether another AVAudioChannelLayout is exactly equal to this layout. @param object The AVAudioChannelLayout to compare against. @discussion The underlying AudioChannelLayoutTag and AudioChannelLayout are compared for equality. */ - (BOOL)isEqual:(id)object; /*! @method layoutWithLayoutTag: @abstract Create from a layout tag. */ + (instancetype)layoutWithLayoutTag:(AudioChannelLayoutTag)layoutTag; /*! @method layoutWithLayout: @abstract Create from an AudioChannelLayout */ + (instancetype)layoutWithLayout:(const AudioChannelLayout *)layout; /*! @property layoutTag @abstract The layout's tag. */ @property (nonatomic, readonly) AudioChannelLayoutTag layoutTag; /*! @property layout @abstract The underlying AudioChannelLayout. */ @property (nonatomic, readonly) const AudioChannelLayout *layout; /*! @property channelCount @abstract The number of channels of audio data. */ @property (nonatomic, readonly) AVAudioChannelCount channelCount; @end NS_ASSUME_NONNULL_END
791
496
# Generated by Django 3.2 on 2021-08-17 23:00 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('posts', '0019_auto_20210528_1122'), ] operations = [ migrations.RemoveField( model_name='historicalpost', name='label', ), migrations.RemoveField( model_name='post', name='label', ), ]
204
634
/**************************************************************** * Licensed to the Apache Software Foundation (ASF) under one * * or more contributor license agreements. See the NOTICE file * * distributed with this work for additional information * * regarding copyright ownership. The ASF licenses this file * * to you under the Apache License, Version 2.0 (the * * "License"); you may not use this file except in compliance * * with the License. You may obtain a copy of the License at * * * * http://www.apache.org/licenses/LICENSE-2.0 * * * * Unless required by applicable law or agreed to in writing, * * software distributed under the License is distributed on an * * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * * KIND, either express or implied. See the License for the * * specific language governing permissions and limitations * * under the License. * ****************************************************************/ package org.apache.james.transport.mailets; import java.util.regex.Pattern; import com.google.common.base.MoreObjects; import com.google.common.base.Objects; public class ReplacingPattern { private final Pattern matcher; private final boolean repeat; private final String substitution; public ReplacingPattern(Pattern matcher, boolean repeat, String substitution) { this.matcher = matcher; this.repeat = repeat; this.substitution = substitution; } public Pattern getMatcher() { return matcher; } public boolean isRepeat() { return repeat; } public String getSubstitution() { return substitution; } @Override public boolean equals(Object o) { if (o instanceof ReplacingPattern) { ReplacingPattern other = (ReplacingPattern) o; return Objects.equal(matcher.pattern(), other.matcher.pattern()) && Objects.equal(repeat, other.repeat) && Objects.equal(substitution, other.substitution); } return false; } @Override public int hashCode() { return Objects.hashCode(matcher, repeat, substitution); } @Override public String toString() { return MoreObjects.toStringHelper(this) .add("matcher", matcher) .add("repeat", repeat) .add("substitution", substitution) .toString(); } }
1,017
5,535
/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. * */ #ifndef _PXFBRIDGE_H #define _PXFBRIDGE_H #include "libchurl.h" #include "pxf_fragment.h" #include "cdb/cdbvars.h" #include "nodes/pg_list.h" /* * Execution state of a foreign scan using pxf_fdw. */ typedef struct PxfFdwScanState { CHURL_HEADERS churl_headers; CHURL_HANDLE churl_handle; StringInfoData uri; ListCell *current_fragment; Relation relation; char *filter_str; ExprState *quals; List *fragments; List *retrieved_attrs; PxfOptions *options; CopyState cstate; } PxfFdwScanState; /* * Execution state of a foreign insert operation. */ typedef struct PxfFdwModifyState { CopyState cstate; /* state of writing to PXF */ CHURL_HANDLE churl_handle; /* curl handle */ CHURL_HEADERS churl_headers; /* curl headers */ StringInfoData uri; /* rest endpoint URI for modify */ Relation relation; PxfOptions *options; /* FDW options */ } PxfFdwModifyState; /* Clean up churl related data structures from the context */ void PxfBridgeCleanup(PxfFdwModifyState *context); /* Sets up data before starting import */ void PxfBridgeImportStart(PxfFdwScanState *pxfsstate); /* Sets up data before starting export */ void PxfBridgeExportStart(PxfFdwModifyState *pxfmstate); /* Reads data from the PXF server into the given buffer of a given size */ int PxfBridgeRead(void *outbuf, int minlen, int maxlen, void *extra); /* Writes data from the given buffer of a given size to the PXF server */ int PxfBridgeWrite(PxfFdwModifyState *context, char *databuf, int datalen); #endif /* _PXFBRIDGE_H */
794
922
# Copyright (c) 2021 - present / Neuralmagic, Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from typing import List import pytest from sparseml.tensorflow_v1.optim import multi_step_lr_schedule, step_lr_schedule from sparseml.tensorflow_v1.utils import tf_compat @pytest.mark.skipif( os.getenv("NM_ML_SKIP_TENSORFLOW_TESTS", False), reason="Skipping tensorflow_v1 tests", ) @pytest.mark.parametrize( "start_step,end_step,init_lr,step_size,gamma", [ (0, 100, 0.01, 1, 0.9), (17, 100, 0.1, 15, 0.9), (0, 200, 0.01, 15, 0.95), (11, 200, 0.1, 7, 0.95), (11, -1, 0.1, 3, 0.95), ], ) def test_step_lr_schedule( start_step: int, end_step: int, init_lr: float, step_size: int, gamma: float ): with tf_compat.Graph().as_default(): global_step = tf_compat.placeholder(dtype=tf_compat.int64, shape=[]) learning_rate = step_lr_schedule( global_step, start_step, end_step, step_size, init_lr, gamma ) with tf_compat.Session() as sess: expected = init_lr for step in range(end_step + 10): measured = sess.run(learning_rate, feed_dict={global_step: step}) if ( step - start_step ) % step_size == 0 and start_step < step <= end_step: expected = expected * gamma assert abs(measured - expected) < 1e-5 @pytest.mark.skipif( os.getenv("NM_ML_SKIP_TENSORFLOW_TESTS", False), reason="Skipping tensorflow_v1 tests", ) @pytest.mark.parametrize( "start_step,milestone_steps,init_lr,gamma", [ (0, [5, 9, 13], 0.01, 0.9), (13, [9, 13], 0.1, 0.9), (0, [5, 9, 13], 0.01, 0.95), (17, [9, 13], 0.1, 0.9), ], ) def test_multi_step_lr_schedule( start_step: int, milestone_steps: List[int], init_lr: float, gamma: float ): with tf_compat.Graph().as_default(): global_step = tf_compat.placeholder(dtype=tf_compat.int64, shape=[]) learning_rate = multi_step_lr_schedule( global_step, start_step, milestone_steps, init_lr, gamma ) with tf_compat.Session() as sess: for step in range(start_step + milestone_steps[-1] + 10): measured = sess.run(learning_rate, feed_dict={global_step: step}) gammas = sum( [1 for mile in milestone_steps if step >= mile + start_step] ) expected = init_lr * gamma ** gammas assert abs(measured - expected) < 1e-5
1,399
335
<reponame>ford442/openmpt<filename>common/mptString.h /* * mptString.h * ---------- * Purpose: Small string-related utilities, number and message formatting. * Notes : Currently none. * Authors: OpenMPT Devs * The OpenMPT source code is released under the BSD license. Read LICENSE for more details. */ #pragma once #include "openmpt/all/BuildSettings.hpp" #include "mpt/base/alloc.hpp" #include "mpt/base/span.hpp" #include "mpt/string/types.hpp" #include "mpt/string/utility.hpp" #include "mptBaseTypes.h" #include <algorithm> #include <limits> #include <string> #include <string_view> #include <cstring> OPENMPT_NAMESPACE_BEGIN namespace mpt { namespace String { template <typename Tstring, typename Tstring2, typename Tstring3> inline Tstring Replace(Tstring str, const Tstring2 &oldStr, const Tstring3 &newStr) { return mpt::replace(str, oldStr, newStr); } } // namespace String enum class Charset { UTF8, ASCII, // strictly 7-bit ASCII ISO8859_1, ISO8859_15, CP850, CP437, CP437AMS, CP437AMS2, Windows1252, #if defined(MPT_ENABLE_CHARSET_LOCALE) Locale, // CP_ACP on windows, current C locale otherwise #endif // MPT_ENABLE_CHARSET_LOCALE }; // source code / preprocessor (i.e. # token) inline constexpr Charset CharsetSource = Charset::ASCII; // debug log files inline constexpr Charset CharsetLogfile = Charset::UTF8; // std::clog / std::cout / std::cerr #if defined(MODPLUG_TRACKER) && MPT_OS_WINDOWS && defined(MPT_ENABLE_CHARSET_LOCALE) inline constexpr Charset CharsetStdIO = Charset::Locale; #else inline constexpr Charset CharsetStdIO = Charset::UTF8; #endif // getenv #if defined(MPT_ENABLE_CHARSET_LOCALE) inline constexpr Charset CharsetEnvironment = Charset::Locale; #else inline constexpr Charset CharsetEnvironment = Charset::UTF8; #endif // std::exception::what() #if defined(MPT_ENABLE_CHARSET_LOCALE) inline constexpr Charset CharsetException = Charset::Locale; #else inline constexpr Charset CharsetException = Charset::UTF8; #endif // Checks if the std::string represents an UTF8 string. // This is currently implemented as converting to std::wstring and back assuming UTF8 both ways, // and comparing the result to the original string. // Caveats: // - can give false negatives because of possible unicode normalization during conversion // - can give false positives if the 8bit encoding contains high-ascii only in valid utf8 groups // - slow because of double conversion bool IsUTF8(const std::string &str); #if MPT_WSTRING_CONVERT // Convert to a wide character string. // The wide encoding is UTF-16 or UTF-32, based on sizeof(wchar_t). // If str does not contain any invalid characters, this conversion is lossless. // Invalid source bytes will be replaced by some replacement character or string. inline std::wstring ToWide(const std::wstring &str) { return str; } inline std::wstring ToWide(const wchar_t * str) { return (str ? std::wstring(str) : std::wstring()); } std::wstring ToWide(Charset from, const std::string &str); inline std::wstring ToWide(Charset from, const char * str) { return ToWide(from, str ? std::string(str) : std::string()); } #if defined(MPT_ENABLE_CHARSET_LOCALE) std::wstring ToWide(const mpt::lstring &str); #endif // MPT_ENABLE_CHARSET_LOCALE #endif // Convert to a string encoded in the 'to'-specified character set. // If str does not contain any invalid characters, // this conversion will be lossless iff, and only iff, // 'to' is UTF8. // Invalid source bytes or characters that are not representable in the // destination charset will be replaced by some replacement character or string. #if MPT_WSTRING_CONVERT std::string ToCharset(Charset to, const std::wstring &str); inline std::string ToCharset(Charset to, const wchar_t * str) { return ToCharset(to, str ? std::wstring(str) : std::wstring()); } #endif std::string ToCharset(Charset to, Charset from, const std::string &str); inline std::string ToCharset(Charset to, Charset from, const char * str) { return ToCharset(to, from, str ? std::string(str) : std::string()); } #if defined(MPT_ENABLE_CHARSET_LOCALE) std::string ToCharset(Charset to, const mpt::lstring &str); #endif // MPT_ENABLE_CHARSET_LOCALE #if defined(MPT_ENABLE_CHARSET_LOCALE) #if MPT_WSTRING_CONVERT mpt::lstring ToLocale(const std::wstring &str); inline mpt::lstring ToLocale(const wchar_t * str) { return ToLocale(str ? std::wstring(str): std::wstring()); } #endif mpt::lstring ToLocale(Charset from, const std::string &str); inline mpt::lstring ToLocale(Charset from, const char * str) { return ToLocale(from, str ? std::string(str): std::string()); } inline mpt::lstring ToLocale(const mpt::lstring &str) { return str; } #endif // MPT_ENABLE_CHARSET_LOCALE #if MPT_OS_WINDOWS #if MPT_WSTRING_CONVERT mpt::winstring ToWin(const std::wstring &str); inline mpt::winstring ToWin(const wchar_t * str) { return ToWin(str ? std::wstring(str): std::wstring()); } #endif mpt::winstring ToWin(Charset from, const std::string &str); inline mpt::winstring ToWin(Charset from, const char * str) { return ToWin(from, str ? std::string(str): std::string()); } #if defined(MPT_ENABLE_CHARSET_LOCALE) mpt::winstring ToWin(const mpt::lstring &str); #endif // MPT_ENABLE_CHARSET_LOCALE #endif // MPT_OS_WINDOWS #if defined(MPT_WITH_MFC) #if !(MPT_WSTRING_CONVERT) #error "MFC depends on MPT_WSTRING_CONVERT)" #endif // Convert to a MFC CString. The CString encoding depends on UNICODE. // This should also be used when converting to TCHAR strings. // If UNICODE is defined, this is a completely lossless operation. inline CString ToCString(const CString &str) { return str; } CString ToCString(const std::wstring &str); inline CString ToCString(const wchar_t * str) { return ToCString(str ? std::wstring(str) : std::wstring()); } CString ToCString(Charset from, const std::string &str); inline CString ToCString(Charset from, const char * str) { return ToCString(from, str ? std::string(str) : std::string()); } #if defined(MPT_ENABLE_CHARSET_LOCALE) CString ToCString(const mpt::lstring &str); mpt::lstring ToLocale(const CString &str); #endif // MPT_ENABLE_CHARSET_LOCALE #if MPT_OS_WINDOWS mpt::winstring ToWin(const CString &str); #endif // MPT_OS_WINDOWS // Convert from a MFC CString. The CString encoding depends on UNICODE. // This should also be used when converting from TCHAR strings. // If UNICODE is defined, this is a completely lossless operation. std::wstring ToWide(const CString &str); std::string ToCharset(Charset to, const CString &str); #endif // MPT_WITH_MFC #define UC_(x) MPT_UCHAR(x) #define UL_(x) MPT_ULITERAL(x) #define U_(x) MPT_USTRING(x) #if MPT_USTRING_MODE_WIDE #if !(MPT_WSTRING_CONVERT) #error "MPT_USTRING_MODE_WIDE depends on MPT_WSTRING_CONVERT)" #endif inline mpt::ustring ToUnicode(const std::wstring &str) { return str; } inline mpt::ustring ToUnicode(const wchar_t * str) { return (str ? std::wstring(str) : std::wstring()); } inline mpt::ustring ToUnicode(Charset from, const std::string &str) { return ToWide(from, str); } inline mpt::ustring ToUnicode(Charset from, const char * str) { return ToUnicode(from, str ? std::string(str) : std::string()); } #if defined(MPT_ENABLE_CHARSET_LOCALE) inline mpt::ustring ToUnicode(const mpt::lstring &str) { return ToWide(str); } #endif // MPT_ENABLE_CHARSET_LOCALE #if defined(MPT_WITH_MFC) inline mpt::ustring ToUnicode(const CString &str) { return ToWide(str); } #endif // MFC #else // !MPT_USTRING_MODE_WIDE inline mpt::ustring ToUnicode(const mpt::ustring &str) { return str; } #if MPT_WSTRING_CONVERT mpt::ustring ToUnicode(const std::wstring &str); inline mpt::ustring ToUnicode(const wchar_t * str) { return ToUnicode(str ? std::wstring(str) : std::wstring()); } #endif mpt::ustring ToUnicode(Charset from, const std::string &str); inline mpt::ustring ToUnicode(Charset from, const char * str) { return ToUnicode(from, str ? std::string(str) : std::string()); } #if defined(MPT_ENABLE_CHARSET_LOCALE) mpt::ustring ToUnicode(const mpt::lstring &str); #endif // MPT_ENABLE_CHARSET_LOCALE #if defined(MPT_WITH_MFC) mpt::ustring ToUnicode(const CString &str); #endif // MPT_WITH_MFC #endif // MPT_USTRING_MODE_WIDE #if MPT_USTRING_MODE_WIDE #if !(MPT_WSTRING_CONVERT) #error "MPT_USTRING_MODE_WIDE depends on MPT_WSTRING_CONVERT)" #endif // nothing, std::wstring overloads will catch all stuff #else // !MPT_USTRING_MODE_WIDE #if MPT_WSTRING_CONVERT std::wstring ToWide(const mpt::ustring &str); #endif std::string ToCharset(Charset to, const mpt::ustring &str); #if defined(MPT_ENABLE_CHARSET_LOCALE) mpt::lstring ToLocale(const mpt::ustring &str); #endif // MPT_ENABLE_CHARSET_LOCALE #if MPT_OS_WINDOWS mpt::winstring ToWin(const mpt::ustring &str); #endif // MPT_OS_WINDOWS #if defined(MPT_WITH_MFC) CString ToCString(const mpt::ustring &str); #endif // MPT_WITH_MFC #endif // MPT_USTRING_MODE_WIDE // The MPT_UTF8 allows specifying UTF8 char arrays. // The resulting type is mpt::ustring and the construction might require runtime translation, // i.e. it is NOT generally available at compile time. // Use explicit UTF8 encoding, // i.e. U+00FC (LATIN SMALL LETTER U WITH DIAERESIS) would be written as "\xC3\xBC". #define MPT_UTF8(x) mpt::ToUnicode(mpt::Charset::UTF8, x) mpt::ustring ToUnicode(uint16 codepage, mpt::Charset fallback, const std::string &str); char ToLowerCaseAscii(char c); char ToUpperCaseAscii(char c); std::string ToLowerCaseAscii(std::string s); std::string ToUpperCaseAscii(std::string s); int CompareNoCaseAscii(const char *a, const char *b, std::size_t n); int CompareNoCaseAscii(std::string_view a, std::string_view b); int CompareNoCaseAscii(const std::string &a, const std::string &b); #if defined(MODPLUG_TRACKER) mpt::ustring ToLowerCase(const mpt::ustring &s); mpt::ustring ToUpperCase(const mpt::ustring &s); #endif // MODPLUG_TRACKER } // namespace mpt // The AnyString types are meant to be used as function argument types only, // and only during the transition phase to all-unicode strings in the whole codebase. // Using an AnyString type as function argument avoids the need to overload a function for all the // different string types that we currently have. // Warning: These types will silently do charset conversions. Only use them when this can be tolerated. // BasicAnyString is convertable to mpt::ustring and constructable from any string at all. template <mpt::Charset charset = mpt::Charset::UTF8, bool tryUTF8 = true> class BasicAnyString : public mpt::ustring { private: static mpt::ustring From8bit(const std::string &str) { if constexpr(charset == mpt::Charset::UTF8) { return mpt::ToUnicode(mpt::Charset::UTF8, str); } else { // auto utf8 detection if constexpr(tryUTF8) { if(mpt::IsUTF8(str)) { return mpt::ToUnicode(mpt::Charset::UTF8, str); } else { return mpt::ToUnicode(charset, str); } } else { return mpt::ToUnicode(charset, str); } } } public: // 8 bit BasicAnyString(const char *str) : mpt::ustring(From8bit(str ? str : std::string())) { } BasicAnyString(const std::string str) : mpt::ustring(From8bit(str)) { } // locale #if defined(MPT_ENABLE_CHARSET_LOCALE) BasicAnyString(const mpt::lstring str) : mpt::ustring(mpt::ToUnicode(str)) { } #endif // MPT_ENABLE_CHARSET_LOCALE // unicode BasicAnyString(const mpt::ustring &str) : mpt::ustring(str) { } BasicAnyString(mpt::ustring &&str) : mpt::ustring(std::move(str)) { } #if MPT_USTRING_MODE_UTF8 && MPT_WSTRING_CONVERT BasicAnyString(const std::wstring &str) : mpt::ustring(mpt::ToUnicode(str)) { } #endif #if MPT_WSTRING_CONVERT BasicAnyString(const wchar_t *str) : mpt::ustring(str ? mpt::ToUnicode(str) : mpt::ustring()) { } #endif // mfc #if defined(MPT_WITH_MFC) BasicAnyString(const CString &str) : mpt::ustring(mpt::ToUnicode(str)) { } #endif // MPT_WITH_MFC // fallback for custom string types template <typename Tstring> BasicAnyString(const Tstring &str) : mpt::ustring(mpt::ToUnicode(str)) { } template <typename Tstring> BasicAnyString(Tstring &&str) : mpt::ustring(mpt::ToUnicode(std::forward<Tstring>(str))) { } }; // AnyUnicodeString is convertable to mpt::ustring and constructable from any unicode string, class AnyUnicodeString : public mpt::ustring { public: // locale #if defined(MPT_ENABLE_CHARSET_LOCALE) AnyUnicodeString(const mpt::lstring &str) : mpt::ustring(mpt::ToUnicode(str)) { } #endif // MPT_ENABLE_CHARSET_LOCALE // unicode AnyUnicodeString(const mpt::ustring &str) : mpt::ustring(str) { } AnyUnicodeString(mpt::ustring &&str) : mpt::ustring(std::move(str)) { } #if MPT_USTRING_MODE_UTF8 && MPT_WSTRING_CONVERT AnyUnicodeString(const std::wstring &str) : mpt::ustring(mpt::ToUnicode(str)) { } #endif #if MPT_WSTRING_CONVERT AnyUnicodeString(const wchar_t *str) : mpt::ustring(str ? mpt::ToUnicode(str) : mpt::ustring()) { } #endif // mfc #if defined(MPT_WITH_MFC) AnyUnicodeString(const CString &str) : mpt::ustring(mpt::ToUnicode(str)) { } #endif // MPT_WITH_MFC // fallback for custom string types template <typename Tstring> AnyUnicodeString(const Tstring &str) : mpt::ustring(mpt::ToUnicode(str)) { } template <typename Tstring> AnyUnicodeString(Tstring &&str) : mpt::ustring(mpt::ToUnicode(std::forward<Tstring>(str))) { } }; // AnyString // Try to do the smartest auto-magic we can do. #if defined(MPT_ENABLE_CHARSET_LOCALE) using AnyString = BasicAnyString<mpt::Charset::Locale, true>; #elif MPT_OS_WINDOWS using AnyString = BasicAnyString<mpt::Charset::Windows1252, true>; #else using AnyString = BasicAnyString<mpt::Charset::ISO8859_1, true>; #endif // AnyStringLocale // char-based strings are assumed to be in locale encoding. #if defined(MPT_ENABLE_CHARSET_LOCALE) using AnyStringLocale = BasicAnyString<mpt::Charset::Locale, false>; #else using AnyStringLocale = BasicAnyString<mpt::Charset::UTF8, false>; #endif // AnyStringUTF8orLocale // char-based strings are tried in UTF8 first, if this fails, locale is used. #if defined(MPT_ENABLE_CHARSET_LOCALE) using AnyStringUTF8orLocale = BasicAnyString<mpt::Charset::Locale, true>; #else using AnyStringUTF8orLocale = BasicAnyString<mpt::Charset::UTF8, false>; #endif // AnyStringUTF8 // char-based strings are assumed to be in UTF8. using AnyStringUTF8 = BasicAnyString<mpt::Charset::UTF8, false>; OPENMPT_NAMESPACE_END
5,344
389
<reponame>tcmoore32/sheer-madness<filename>gosu-test-api/src/main/java/gw/test/BaseRemoteTestClass.java<gh_stars>100-1000 package gw.test; /* * Copyright 2014 Guidewire Software, Inc. */ public class BaseRemoteTestClass extends TestClass { protected BaseRemoteTestClass(String s, boolean shouldInit) { super(s, shouldInit); } }
119
570
<reponame>AmbientAI/amazon-kinesis-client<gh_stars>100-1000 /* * Copyright 2019 Amazon.com, Inc. or its affiliates. * Licensed under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package software.amazon.kinesis.leases; import java.util.ArrayList; import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Map; /** * Shard Prioritization that prioritizes parent shards first. * It also limits number of shards that will be available for initialization based on their depth. * It doesn't make a lot of sense to work on a shard that has too many unfinished parents. */ public class ParentsFirstShardPrioritization implements ShardPrioritization { private static final SortingNode PROCESSING_NODE = new SortingNode(null, Integer.MIN_VALUE); private final int maxDepth; /** * Creates ParentFirst prioritization with filtering based on depth of the shard. * Shards that have depth > maxDepth will be ignored and will not be returned by this prioritization. * * @param maxDepth any shard that is deeper than max depth, will be excluded from processing */ public ParentsFirstShardPrioritization(int maxDepth) { /* Depth 0 means that shard is completed or cannot be found, * it is impossible to process such shards. */ if (maxDepth <= 0) { throw new IllegalArgumentException("Max depth cannot be negative or zero. Provided value: " + maxDepth); } this.maxDepth = maxDepth; } @Override public List<ShardInfo> prioritize(List<ShardInfo> original) { Map<String, ShardInfo> shards = new HashMap<>(); for (ShardInfo shardInfo : original) { shards.put(shardInfo.shardId(), shardInfo); } Map<String, SortingNode> processedNodes = new HashMap<>(); for (ShardInfo shardInfo : original) { populateDepth(shardInfo.shardId(), shards, processedNodes); } List<ShardInfo> orderedInfos = new ArrayList<>(original.size()); List<SortingNode> orderedNodes = new ArrayList<>(processedNodes.values()); Collections.sort(orderedNodes); for (SortingNode sortingTreeNode : orderedNodes) { // don't process shards with depth > maxDepth if (sortingTreeNode.getDepth() <= maxDepth) { orderedInfos.add(sortingTreeNode.shardInfo); } } return orderedInfos; } private int populateDepth(String shardId, Map<String, ShardInfo> shards, Map<String, SortingNode> processedNodes) { SortingNode processed = processedNodes.get(shardId); if (processed != null) { if (processed == PROCESSING_NODE) { throw new IllegalArgumentException("Circular dependency detected. Shard Id " + shardId + " is processed twice"); } return processed.getDepth(); } ShardInfo shardInfo = shards.get(shardId); if (shardInfo == null) { // parent doesn't exist in our list, so this shard is root-level node return 0; } if (shardInfo.isCompleted()) { // we treat completed shards as 0-level return 0; } // storing processing node to make sure we track progress and avoid circular dependencies processedNodes.put(shardId, PROCESSING_NODE); int maxParentDepth = 0; for (String parentId : shardInfo.parentShardIds()) { maxParentDepth = Math.max(maxParentDepth, populateDepth(parentId, shards, processedNodes)); } int currentNodeLevel = maxParentDepth + 1; SortingNode previousValue = processedNodes.put(shardId, new SortingNode(shardInfo, currentNodeLevel)); if (previousValue != PROCESSING_NODE) { throw new IllegalStateException("Validation failed. Depth for shardId " + shardId + " was populated twice"); } return currentNodeLevel; } /** * Class to store depth of shards during prioritization. */ private static class SortingNode implements Comparable<SortingNode> { private final ShardInfo shardInfo; private final int depth; public SortingNode(ShardInfo shardInfo, int depth) { this.shardInfo = shardInfo; this.depth = depth; } public int getDepth() { return depth; } @Override public int compareTo(SortingNode o) { return Integer.compare(depth, o.depth); } } }
2,145
315
<reponame>FredrikBlomgren/aff3ct<filename>include/Tools/Algo/Draw_generator/Event_generator/MKL/Event_generator_MKL.hpp /*! * \file * \brief Class tools::Event_generator_MKL. */ #ifdef AFF3CT_CHANNEL_MKL #ifndef EVENT_GENERATOR_MKL_HPP #define EVENT_GENERATOR_MKL_HPP #include "Tools/types.h" #include "Tools/Algo/Draw_generator/Event_generator/Event_generator.hpp" namespace aff3ct { namespace tools { template <typename R = float, typename E = typename tools::matching_types<R>::B> class Event_generator_MKL : public Event_generator<R,E> { private: void* stream_state; // VSLStreamStatePtr* type bool is_stream_alloc; public: explicit Event_generator_MKL(const int seed = 0); virtual ~Event_generator_MKL(); virtual Event_generator_MKL<R,E>* clone() const; virtual void set_seed(const int seed); virtual void generate(E *draw, const unsigned length, const R event_probability); }; } } #endif //EVENT_GENERATOR_MKL_HPP #endif // MKL
370
575
// Copyright (c) 2011 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "printing/printed_document.h" #include "base/check_op.h" #include "base/strings/utf_string_conversions.h" #include "printing/metafile_skia.h" #include "printing/page_number.h" #include "printing/printed_page_win.h" #include "printing/printing_context_win.h" #include "printing/units.h" #include "skia/ext/skia_utils_win.h" namespace { void SimpleModifyWorldTransform(HDC context, int offset_x, int offset_y, float shrink_factor) { XFORM xform = {0}; xform.eDx = static_cast<float>(offset_x); xform.eDy = static_cast<float>(offset_y); xform.eM11 = xform.eM22 = 1.f / shrink_factor; BOOL res = ModifyWorldTransform(context, &xform, MWT_LEFTMULTIPLY); DCHECK_NE(res, 0); } } // namespace namespace printing { void PrintedDocument::RenderPrintedPage( const PrintedPage& page, printing::NativeDrawingContext context) const { #ifndef NDEBUG { // Make sure the page is from our list. base::AutoLock lock(lock_); DCHECK(&page == mutable_.pages_.find(page.page_number() - 1)->second.get()); } #endif DCHECK(context); const PageSetup& page_setup = immutable_.settings_->page_setup_device_units(); gfx::Rect content_area = GetCenteredPageContentRect( page_setup.physical_size(), page.page_size(), page.page_content_rect()); // Save the state to make sure the context this function call does not modify // the device context. int saved_state = SaveDC(context); DCHECK_NE(saved_state, 0); skia::InitializeDC(context); { // Save the state (again) to apply the necessary world transformation. int saved_state_inner = SaveDC(context); DCHECK_NE(saved_state_inner, 0); // Setup the matrix to translate and scale to the right place. Take in // account the actual shrinking factor. // Note that the printing output is relative to printable area of the page. // That is 0,0 is offset by PHYSICALOFFSETX/Y from the page. SimpleModifyWorldTransform( context, content_area.x() - page_setup.printable_area().x(), content_area.y() - page_setup.printable_area().y(), page.shrink_factor()); ::StartPage(context); bool played_back = page.metafile()->SafePlayback(context); DCHECK(played_back); ::EndPage(context); BOOL res = RestoreDC(context, saved_state_inner); DCHECK_NE(res, 0); } BOOL res = RestoreDC(context, saved_state); DCHECK_NE(res, 0); } bool PrintedDocument::RenderPrintedDocument(PrintingContext* context) { if (context->NewPage() != PrintingContext::OK) return false; std::wstring device_name = base::UTF16ToWide(immutable_.settings_->device_name()); { base::AutoLock lock(lock_); const MetafilePlayer* metafile = GetMetafile(); static_cast<PrintingContextWin*>(context)->PrintDocument( device_name, *(static_cast<const MetafileSkia*>(metafile))); } return context->PageDone() == PrintingContext::OK; } } // namespace printing
1,185
2,083
/* * Copyright (C) 2015 Karumi. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package com.karumi.rosie.repository.datasource.paginated; import com.karumi.rosie.repository.PaginatedCollection; import com.karumi.rosie.repository.datasource.Identifiable; import java.util.Collection; /** * Data source interface meant to be used only to persist page of data. * * @param <K> The class of the key used by this data source. * @param <V> The class of the values stored into this data source. */ public interface PaginatedWriteableDataSource<K, V extends Identifiable<K>> { /** * Adds or update a page of values into this data source. * * @param page Page to be stored * @param values Collection of values to be stored * @param hasMore True whether the persisted page has more elements */ PaginatedCollection<V> addOrUpdatePage(Page page, Collection<V> values, boolean hasMore) throws Exception; /** * Deletes all the pages stored in this data source. */ void deleteAll() throws Exception; }
438
386
<reponame>beru/libonnx<filename>src/default/Mean.c<gh_stars>100-1000 #include <onnx.h> static int Mean_init(struct onnx_node_t * n) { if((n->ninput >= 1) && (n->noutput == 1)) return 1; return 0; } static int Mean_exit(struct onnx_node_t * n) { return 1; } static int Mean_reshape(struct onnx_node_t * n) { struct onnx_tensor_t * y = n->outputs[0]; int i; if(!onnx_tensor_reshape_identity(y, n->inputs[0], n->inputs[0]->type)) return 0; for(i = 1; i < n->ninput; i++) { if(!onnx_tensor_reshape_multi_broadcast(y, y, n->inputs[i], y->type)) return 0; } return 1; } static void Mean_bfloat16(struct onnx_node_t * n) { struct onnx_tensor_t * y = n->outputs[0]; struct onnx_tensor_t * x; uint16_t * py = (uint16_t *)y->datas; uint16_t * px; float sum; size_t i, j, l; for(i = 0, l = y->ndata; i < l; i++) { for(j = 0, sum = 0; j < n->ninput; j++) { x = n->inputs[j]; px = onnx_tensor_broadcast_map_address(x, y, i); sum += bfloat16_to_float32(*px); } py[i] = float32_to_bfloat16(sum / n->ninput); } } static void Mean_float16(struct onnx_node_t * n) { struct onnx_tensor_t * y = n->outputs[0]; struct onnx_tensor_t * x; uint16_t * py = (uint16_t *)y->datas; uint16_t * px; float sum; size_t i, j, l; for(i = 0, l = y->ndata; i < l; i++) { for(j = 0, sum = 0; j < n->ninput; j++) { x = n->inputs[j]; px = onnx_tensor_broadcast_map_address(x, y, i); sum += float16_to_float32(*px); } py[i] = float32_to_float16(sum / n->ninput); } } static void Mean_float32(struct onnx_node_t * n) { struct onnx_tensor_t * y = n->outputs[0]; struct onnx_tensor_t * x; float * py = (float *)y->datas; float * px; float sum; size_t i, j, l; for(i = 0, l = y->ndata; i < l; i++) { for(j = 0, sum = 0; j < n->ninput; j++) { x = n->inputs[j]; px = onnx_tensor_broadcast_map_address(x, y, i); sum += *px; } py[i] = sum / n->ninput; } } static void Mean_float64(struct onnx_node_t * n) { struct onnx_tensor_t * y = n->outputs[0]; struct onnx_tensor_t * x; double * py = (double *)y->datas; double * px; double sum; size_t i, j, l; for(i = 0, l = y->ndata; i < l; i++) { for(j = 0, sum = 0; j < n->ninput; j++) { x = n->inputs[j]; px = onnx_tensor_broadcast_map_address(x, y, i); sum += *px; } py[i] = sum / n->ninput; } } void resolver_default_op_Mean(struct onnx_node_t * n) { if(n->opset >= 13) { switch(n->inputs[0]->type) { case ONNX_TENSOR_TYPE_BFLOAT16: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_bfloat16; break; case ONNX_TENSOR_TYPE_FLOAT16: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float16; break; case ONNX_TENSOR_TYPE_FLOAT32: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float32; break; case ONNX_TENSOR_TYPE_FLOAT64: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float64; break; default: break; } } else if(n->opset >= 8) { switch(n->inputs[0]->type) { case ONNX_TENSOR_TYPE_FLOAT16: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float16; break; case ONNX_TENSOR_TYPE_FLOAT32: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float32; break; case ONNX_TENSOR_TYPE_FLOAT64: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float64; break; default: break; } } else if(n->opset >= 6) { switch(n->inputs[0]->type) { case ONNX_TENSOR_TYPE_FLOAT16: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float16; break; case ONNX_TENSOR_TYPE_FLOAT32: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float32; break; case ONNX_TENSOR_TYPE_FLOAT64: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float64; break; default: break; } } else if(n->opset >= 1) { switch(n->inputs[0]->type) { case ONNX_TENSOR_TYPE_FLOAT16: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float16; break; case ONNX_TENSOR_TYPE_FLOAT32: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float32; break; case ONNX_TENSOR_TYPE_FLOAT64: n->init = Mean_init; n->exit = Mean_exit; n->reshape = Mean_reshape; n->operator = Mean_float64; break; default: break; } } }
2,428
1,405
package a.a.a.a.a.a; import java.io.InputStream; import java.io.OutputStream; public final class j implements o { /* renamed from: a reason: collision with root package name */ private Class f20a; private String b; private Object c; public j(String str) { this.b = str; } @Override // a.a.a.a.a.a.o public final void a() { if (!h.a("com.ibm.mqttdirect.modules.local.bindings.LocalListener")) { throw h.a(32103); } try { this.f20a = Class.forName("com.ibm.mqttdirect.modules.local.bindings.LocalListener"); this.c = this.f20a.getMethod("connect", String.class).invoke(null, this.b); } catch (Exception e) { } if (this.c == null) { throw h.a(32103); } } @Override // a.a.a.a.a.a.o public final InputStream b() { try { return (InputStream) this.f20a.getMethod("getClientInputStream", new Class[0]).invoke(this.c, new Object[0]); } catch (Exception e) { return null; } } @Override // a.a.a.a.a.a.o public final OutputStream c() { try { return (OutputStream) this.f20a.getMethod("getClientOutputStream", new Class[0]).invoke(this.c, new Object[0]); } catch (Exception e) { return null; } } @Override // a.a.a.a.a.a.o public final void d() { if (this.c != null) { try { this.f20a.getMethod("close", new Class[0]).invoke(this.c, new Object[0]); } catch (Exception e) { } } } }
793
764
<reponame>641589523/token-profile<filename>erc20/0x15F0EEDF9Ce24fc4b6826E590A8292CE5524a1DA.json {"symbol": "DENA","address": "0x15F0EEDF9Ce24fc4b6826E590A8292CE5524a1DA","overview":{"en": ""},"email": "","website": "https://denations.com/","state": "NORMAL","links": {"blog": "https://smatoos.medium.com/","twitter": "https://twitter.com/SMATOOS_now","telegram": "https://t.me/smatoos","github": "https://github.com/denation-smatoos/denations"}}
184
2,045
''' This is the benchmark of 4 different implementations of rectified linear activation in Theano. Two types of computations are tested w.r.t. each implementation: fprop and grad. Results: in seconds, float32 (details in the code) Implementation, CPU (fprop, bprop), GPU (fprop, bprop), (final score) a) ScalarRectifier: (2.32, 2.40) (1.36, 2.67) (8.75) b) T.max(.0, x): (5.19, 3.65) (1.38, 2.38) (12.60) c) x*(x>0.): (2.85, 2.84) (1.31, 2.91) (9.91) d) T.switch(x<0., 0., x): (2.32, 1.41) (1.41, 2.84) (8.39) Conlusion: In terms of efficiency, d) > a) > c) > b) ''' from __future__ import print_function __authors__ = "<NAME> and <NAME>" import theano import theano.tensor as T import numpy import time floatX = 'float32' def relu(x): """ relu implementation with T.maximum Parameters ---------- x: tensor variable """ return T.maximum(0.0, x) def relu_(x): """ Alternative relu implementation Parameters ---------- x: tensor variable """ return x * (x > 0) def relu__(x): """ Alternative relu implementation. The most efficient one. Parameters ---------- x: tensor variable """ return T.switch(x < 0., 0., x) def test_scalar_rectifier(): """ verify different implementations of relu """ x = T.fmatrix('inputs') y1 = relu(x) y3 = relu_(x) y4 = relu__(x) f1 = theano.function(inputs=[x], outputs=y1, name='benchmark_1_forward') f3 = theano.function(inputs=[x], outputs=y3, name='benchmark_3_forward') f4 = theano.function(inputs=[x], outputs=y4, name='benchmark_4_forward') g1 = theano.function(inputs=[x], outputs=T.grad(y1.sum(),x), name='benchmark_1_grad') g3 = theano.function(inputs=[x], outputs=T.grad(y3.sum(),x), name='benchmark_3_grad') g4 = theano.function(inputs=[x], outputs=T.grad(y4.sum(),x), name='benchmark_4_grad') for i in range(10): value = numpy.random.uniform(-1,1,size=(100,500)).astype(floatX) numpy.testing.assert_array_equal(f1(value), f3(value), err_msg='arrays not equal' ) numpy.testing.assert_array_equal(f1(value), f4(value), err_msg='arrays not equal' ) numpy.testing.assert_array_equal(g1(value), g3(value), err_msg='grad:arrays not equal' ) numpy.testing.assert_array_equal(g1(value), g4(value), err_msg='grad:arrays not equal' ) def benchmark_relu(): """ Benchmark the speed of different relu implementations. Both fprop and grad are tested. """ x = T.ftensor4('inputs') ops = [ relu_(x).sum(), # old relu(x).sum(), # alter, short for alternative relu__(x).sum(), # alter 2 T.grad(relu_(x).sum(),x), # grad_old T.grad(relu(x).sum(),x), # grad_alter T.grad(relu__(x).sum(),x), # grad_alter2 ] names = ['fprop_old', 'fprop_alter', 'fprop_alter2', 'grad_old', 'grad_alter', 'grad_alter2'] value = numpy.random.uniform(size=(512,32,32,100)).astype(floatX) times = [] for op, name in zip(ops, names): f = theano.function(inputs=[x], outputs=op, name=name) n_loops = 10 t0 = time.time() for i in range(n_loops): f(value) t1 = time.time() benchmark = t1-t0 times.append(benchmark) print(name) theano.printing.debugprint(f, print_type=True) print(names) print(times) if __name__ == '__main__': benchmark_relu() #test_scalar_rectifier()
1,780
649
package net.thucydides.core.reports.integration; import net.thucydides.core.annotations.Feature; import net.thucydides.core.annotations.Story; import net.thucydides.core.issues.IssueTracking; import net.thucydides.core.issues.SystemPropertiesIssueTracking; import net.thucydides.core.model.TestOutcome; import net.thucydides.core.model.TestStep; import net.thucydides.core.reports.AcceptanceTestReporter; import net.thucydides.core.reports.html.HtmlAcceptanceTestReporter; import net.thucydides.core.screenshots.ScreenshotAndHtmlSource; import net.thucydides.core.util.ExtendedTemporaryFolder; import net.thucydides.core.util.FileSystemUtils; import net.thucydides.core.util.MockEnvironmentVariables; import org.apache.commons.io.FileUtils; import org.junit.Before; import org.junit.Rule; import java.io.File; import java.io.IOException; public class AbstractReportGenerationTest { @Rule public ExtendedTemporaryFolder temporaryDirectory = new ExtendedTemporaryFolder(); protected AcceptanceTestReporter reporter; protected File outputDirectory; protected MockEnvironmentVariables environmentVariables; public AbstractReportGenerationTest() { } @Before public void setupTestReporter() throws IOException { environmentVariables = new MockEnvironmentVariables(); IssueTracking issueTracking = new SystemPropertiesIssueTracking(environmentVariables); reporter = new HtmlAcceptanceTestReporter(environmentVariables, issueTracking); outputDirectory = temporaryDirectory.newFolder(); reporter.setOutputDirectory(outputDirectory); } protected void recordStepWithScreenshot(final TestOutcome testOutcome, final String stepName, final String screenshot) throws IOException { String screenshotResource = "/screenshots/" + screenshot; File sourceFile = FileSystemUtils.getResourceAsFile(screenshotResource);// new File(sourcePath.getPath()); FileUtils.copyFileToDirectory(sourceFile, outputDirectory); TestStep step = TestStepFactory.successfulTestStepCalled(stepName); step.addScreenshot(new ScreenshotAndHtmlSource(new File(outputDirectory, screenshot), null)); testOutcome.recordStep(step); } protected class AUserStory { } @Story(WhenGeneratingAnHtmlReport.AUserStory.class) protected class SomeTestScenario { public void a_simple_test_case() { } ; public void should_do_this() { } ; public void should_do_that() { } ; } @Feature protected class AFeature { class AUserStoryInAFeature { } ; } @Story(WhenGeneratingAnHtmlReport.AFeature.AUserStoryInAFeature.class) protected class SomeTestScenarioInAFeature { public void should_do_this() { } ; public void should_do_that() { } ; } }
1,129
357
<reponame>nicesu/wiki<filename>Programming/Java/Collection/StackDemo.java<gh_stars>100-1000 import java.util.*; public class StackDemo { public static void main(String args[]) { Stack s = new Stack(); s.push(new Integer(0)); s.push(new Student(202)); s.push(new Integer(1)); //push in System.out.println( (Integer)s.peek()); //get but not pop System.out.println((Integer)s.pop()); //pop out Student a =(Student)s.pop(); a.prt(); System.out.println((Integer)s.pop()); } } class Student { int no; Student(int i) {no = i;} void prt(){System.out.println(no);} }
275
666
from lib import BaseTest class PublishShow1Test(BaseTest): """ publish show: existing snapshot """ fixtureDB = True fixturePool = True fixtureCmds = [ "aptly snapshot create snap1 from mirror gnuplot-maverick", "aptly publish snapshot -keyring=${files}/aptly.pub -secret-keyring=${files}/aptly.sec snap1", ] runCmd = "aptly publish show maverick" class PublishShow2Test(BaseTest): """ publish show: under prefix """ fixtureDB = True fixturePool = True fixtureCmds = [ "aptly snapshot create snap1 from mirror gnuplot-maverick", "aptly publish snapshot -keyring=${files}/aptly.pub -secret-keyring=${files}/aptly.sec snap1 ppa/smira", ] runCmd = "aptly publish show maverick ppa/smira" class PublishShow3Test(BaseTest): """ publish show json: existing snapshot """ fixtureDB = True fixturePool = True fixtureCmds = [ "aptly snapshot create snap1 from mirror gnuplot-maverick", "aptly publish snapshot -keyring=${files}/aptly.pub -secret-keyring=${files}/aptly.sec snap1", ] runCmd = "aptly publish show -json maverick" class PublishShow4Test(BaseTest): """ publish show json: under prefix """ fixtureDB = True fixturePool = True fixtureCmds = [ "aptly snapshot create snap1 from mirror gnuplot-maverick", "aptly publish snapshot -keyring=${files}/aptly.pub -secret-keyring=${files}/aptly.sec snap1 ppa/smira", ] runCmd = "aptly publish show -json maverick ppa/smira"
608
331
<filename>Easy Challenges/Challenge 0020 Easy/solutions/solution.java package dailyprogrammerchallenges; import static java.lang.Math.sqrt; import java.util.Scanner; /** * * @author coderMato */ public class PrimeNumbers { public static void main(String[] args){ int prime,i,j; System.out.println("2"); for(i=3; i <= 2000; i++){ for(j=2; j <= (int) sqrt(i) + 1; j++){ if(i % j != 0 && j == (int) sqrt(i) + 1){ System.out.println(i); } if (i%j == 0){ break; } } } } }
344
13,057
<filename>src/main/java/org/mockito/plugins/InlineMockMaker.java /* * Copyright (c) 2019 Mockito contributors * This program is made available under the terms of the MIT License. */ package org.mockito.plugins; import org.mockito.MockitoFramework; /** * Extension to {@link MockMaker} for mock makers that changes inline method implementations * and need keep track of created mock objects. * <p> * Mockito's default inline mock maker keeps track of created mock objects via weak reference map. * This poses a risk of memory leaks in certain scenarios * (issue <a href="https://github.com/mockito/mockito/pull/1619">#1619</a>). * There is no clean way to tackle those problems at the moment. * Hence, {@code InlineMockMaker} interface exposes methods to explicitly clear mock references. * Those methods are called by {@link MockitoFramework#clearInlineMocks()}. * When the user encounters a leak, he can mitigate the problem with {@link MockitoFramework#clearInlineMocks()}. * <p> * {@code InlineMockMaker} is for expert users and framework integrators, when custom inline mock maker is in use. * If you have a custom {@link MockMaker} that keeps track of mock objects, * please have your mock maker implement {@code InlineMockMaker} interface. * This way, it can participate in {@link MockitoFramework#clearInlineMocks()} API. * * @since 2.25.0 */ public interface InlineMockMaker extends MockMaker { /** * Clean up internal state for specified {@code mock}. You may assume there won't be any interaction to the specific * mock after this is called. * * @param mock the mock instance whose internal state is to be cleaned. * @since 2.25.0 */ void clearMock(Object mock); /** * Cleans up internal state for all existing mocks. You may assume there won't be any interaction to mocks created * previously after this is called. * * @since 2.25.0 */ void clearAllMocks(); }
582
1,958
package com.freetymekiyan.algorithms.level.hard; /** * The demons had captured the princess (P) and imprisoned her in the * bottom-right corner of a dungeon. The dungeon consists of M x N rooms laid * out in a 2D grid. Our valiant knight (K) was initially positioned in the * top-left room and must fight his way through the dungeon to rescue the * princess. * <p> * The knight has an initial health point represented by a positive integer. If * at any point his health point drops to 0 or below, he dies immediately. * <p> * Some of the rooms are guarded by demons, so the knight loses health * (negative integers) upon entering these rooms; other rooms are either empty * (0's) or contain magic orbs that increase the knight's health (positive * integers). * <p> * In order to reach the princess as quickly as possible, the knight decides to * move only rightward or downward in each step. * <p> * Write a function to determine the knight's minimum initial health so that he * is able to rescue the princess. * <p> * For example, given the dungeon below, the initial health of the knight must * be at least 7 if he follows the optimal path RIGHT-> RIGHT -> DOWN -> DOWN. * <p> * -2(K) -3 3 * -5 -10 1 * 10 30 -5(P) * <p> * Notes: * The knight's health has no upper bound. * Any room can contain threats or power-ups, even the first room the knight * enters and the bottom-right room where the princess is imprisoned. * <p> * Tags:DP, Binary Search */ class DungeonGame { /** * Build a table for the minimum hp needed to get to the bottom right * Build from bottom right to get minimum from i, j to the end * Instead of build from to-left, because it's hard to get correct relation */ public int calculateMinimumHP(int[][] dungeon) { if (dungeon == null || dungeon.length == 0 || dungeon[0].length == 0) return 0; int m = dungeon.length - 1; int n = dungeon[0].length - 1; dungeon[m][n] = Math.max(1 - dungeon[m][n], 1); for (int i = m - 1; i >= 0; i--) dungeon[i][n] = Math.max(dungeon[i + 1][n] - dungeon[i][n], 1); for (int j = n - 1; j >= 0; j--) dungeon[m][j] = Math.max(dungeon[m][j + 1] - dungeon[m][j], 1); for (int i = m - 1; i >= 0; i--) { for (int j = n - 1; j >= 0; j--) { dungeon[i][j] = Math.max(Math.min(dungeon[i + 1][j], dungeon[i][j + 1]) - dungeon[i][j], 1); } } return dungeon[0][0]; } /** * Why “from the bottom right corner to left top”? * It depends on the way you formulate the problem. If you define a value * in DP table d[i][j] as 'the minimum hp required to REACH (i, j) from (0, * 0)", then the final answer should be d[nrows-1][ncols-1], and you need * to start filling from the top left; * However, in the reference answer provided with the question, dp[i][j] is * defined as 'the minimum hp required to REACH (nrows-1, ncols-1) from (i, * j)'. Here dp[0][0] is the final answer so we must fill from (nrows-1, * ncols-1). For many other problems such as 'Minimum Path Sum', both * formulation would work. * However, in this problem, the former formulation will lead us to * trouble, because it is very hard, if not impossible, to get d[i][j] * based on d[i-1][j] and d[i][j-1]. */ }
1,107
2,338
// RUN: %libomp-compile-and-run #include <stdio.h> #include <math.h> #include "omp_testsuite.h" #include "omp_my_sleep.h" int test_omp_task_final() { int tids[NUM_TASKS]; int includedtids[NUM_TASKS]; int i; int error = 0; #pragma omp parallel { #pragma omp single { for (i = 0; i < NUM_TASKS; i++) { /* First we have to store the value of the loop index in a new variable * which will be private for each task because otherwise it will be overwritten * if the execution of the task takes longer than the time which is needed to * enter the next step of the loop! */ int myi; myi = i; #pragma omp task final(i>=10) { tids[myi] = omp_get_thread_num(); /* we generate included tasks for final tasks */ if(myi >= 10) { int included = myi; #pragma omp task { my_sleep (SLEEPTIME); includedtids[included] = omp_get_thread_num(); } /* end of omp included task of the final task */ my_sleep (SLEEPTIME); } /* end of if it is a final task*/ } /* end of omp task */ } /* end of for */ } /* end of single */ } /*end of parallel */ /* Now we ckeck if more than one thread executed the final task and its included task. */ for (i = 10; i < NUM_TASKS; i++) { if (tids[i] != includedtids[i]) { error++; } } return (error==0); } /* end of check_paralel_for_private */ int main() { int i; int num_failed=0; for(i = 0; i < REPETITIONS; i++) { if(!test_omp_task_final()) { num_failed++; } } return num_failed; }
775
777
<gh_stars>100-1000 // Copyright 2016 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "core/animation/CSSImageSliceInterpolationType.h" #include "core/animation/CSSLengthInterpolationType.h" #include "core/animation/ImageSlicePropertyFunctions.h" #include "core/css/CSSBorderImageSliceValue.h" #include "core/css/resolver/StyleResolverState.h" #include "wtf/PtrUtil.h" #include <memory> namespace blink { namespace { enum SideIndex : unsigned { SideTop, SideRight, SideBottom, SideLeft, SideIndexCount, }; struct SliceTypes { explicit SliceTypes(const ImageSlice& slice) { isNumber[SideTop] = slice.slices.top().isFixed(); isNumber[SideRight] = slice.slices.right().isFixed(); isNumber[SideBottom] = slice.slices.bottom().isFixed(); isNumber[SideLeft] = slice.slices.left().isFixed(); fill = slice.fill; } explicit SliceTypes(const CSSBorderImageSliceValue& slice) { isNumber[SideTop] = slice.slices().top()->isPrimitiveValue() && toCSSPrimitiveValue(slice.slices().top())->isNumber(); isNumber[SideRight] = slice.slices().right()->isPrimitiveValue() && toCSSPrimitiveValue(slice.slices().right())->isNumber(); isNumber[SideBottom] = slice.slices().bottom()->isPrimitiveValue() && toCSSPrimitiveValue(slice.slices().bottom())->isNumber(); isNumber[SideLeft] = slice.slices().left()->isPrimitiveValue() && toCSSPrimitiveValue(slice.slices().left())->isNumber(); fill = slice.fill(); } bool operator==(const SliceTypes& other) const { for (size_t i = 0; i < SideIndexCount; i++) { if (isNumber[i] != other.isNumber[i]) return false; } return fill == other.fill; } bool operator!=(const SliceTypes& other) const { return !(*this == other); } // If a side is not a number then it is a percentage. bool isNumber[SideIndexCount]; bool fill; }; } // namespace class CSSImageSliceNonInterpolableValue : public NonInterpolableValue { public: static PassRefPtr<CSSImageSliceNonInterpolableValue> create( const SliceTypes& types) { return adoptRef(new CSSImageSliceNonInterpolableValue(types)); } const SliceTypes& types() const { return m_types; } DECLARE_NON_INTERPOLABLE_VALUE_TYPE(); private: CSSImageSliceNonInterpolableValue(const SliceTypes& types) : m_types(types) {} const SliceTypes m_types; }; DEFINE_NON_INTERPOLABLE_VALUE_TYPE(CSSImageSliceNonInterpolableValue); DEFINE_NON_INTERPOLABLE_VALUE_TYPE_CASTS(CSSImageSliceNonInterpolableValue); namespace { class UnderlyingSliceTypesChecker : public InterpolationType::ConversionChecker { public: static std::unique_ptr<UnderlyingSliceTypesChecker> create( const SliceTypes& underlyingTypes) { return WTF::wrapUnique(new UnderlyingSliceTypesChecker(underlyingTypes)); } static SliceTypes getUnderlyingSliceTypes( const InterpolationValue& underlying) { return toCSSImageSliceNonInterpolableValue(*underlying.nonInterpolableValue) .types(); } private: UnderlyingSliceTypesChecker(const SliceTypes& underlyingTypes) : m_underlyingTypes(underlyingTypes) {} bool isValid(const InterpolationEnvironment&, const InterpolationValue& underlying) const final { return m_underlyingTypes == getUnderlyingSliceTypes(underlying); } const SliceTypes m_underlyingTypes; }; class InheritedSliceTypesChecker : public InterpolationType::ConversionChecker { public: static std::unique_ptr<InheritedSliceTypesChecker> create( CSSPropertyID property, const SliceTypes& inheritedTypes) { return WTF::wrapUnique( new InheritedSliceTypesChecker(property, inheritedTypes)); } private: InheritedSliceTypesChecker(CSSPropertyID property, const SliceTypes& inheritedTypes) : m_property(property), m_inheritedTypes(inheritedTypes) {} bool isValid(const InterpolationEnvironment& environment, const InterpolationValue& underlying) const final { return m_inheritedTypes == SliceTypes(ImageSlicePropertyFunctions::getImageSlice( m_property, *environment.state().parentStyle())); } const CSSPropertyID m_property; const SliceTypes m_inheritedTypes; }; InterpolationValue convertImageSlice(const ImageSlice& slice, double zoom) { std::unique_ptr<InterpolableList> list = InterpolableList::create(SideIndexCount); const Length* sides[SideIndexCount] = {}; sides[SideTop] = &slice.slices.top(); sides[SideRight] = &slice.slices.right(); sides[SideBottom] = &slice.slices.bottom(); sides[SideLeft] = &slice.slices.left(); for (size_t i = 0; i < SideIndexCount; i++) { const Length& side = *sides[i]; list->set(i, InterpolableNumber::create( side.isFixed() ? side.pixels() / zoom : side.percent())); } return InterpolationValue( std::move(list), CSSImageSliceNonInterpolableValue::create(SliceTypes(slice))); } } // namespace InterpolationValue CSSImageSliceInterpolationType::maybeConvertNeutral( const InterpolationValue& underlying, ConversionCheckers& conversionCheckers) const { SliceTypes underlyingTypes = UnderlyingSliceTypesChecker::getUnderlyingSliceTypes(underlying); conversionCheckers.push_back( UnderlyingSliceTypesChecker::create(underlyingTypes)); LengthBox zeroBox( Length(0, underlyingTypes.isNumber[SideTop] ? Fixed : Percent), Length(0, underlyingTypes.isNumber[SideRight] ? Fixed : Percent), Length(0, underlyingTypes.isNumber[SideBottom] ? Fixed : Percent), Length(0, underlyingTypes.isNumber[SideLeft] ? Fixed : Percent)); return convertImageSlice(ImageSlice(zeroBox, underlyingTypes.fill), 1); } InterpolationValue CSSImageSliceInterpolationType::maybeConvertInitial( const StyleResolverState&, ConversionCheckers& conversionCheckers) const { return convertImageSlice( ImageSlicePropertyFunctions::getInitialImageSlice(cssProperty()), 1); } InterpolationValue CSSImageSliceInterpolationType::maybeConvertInherit( const StyleResolverState& state, ConversionCheckers& conversionCheckers) const { const ImageSlice& inheritedImageSlice = ImageSlicePropertyFunctions::getImageSlice(cssProperty(), *state.parentStyle()); conversionCheckers.push_back(InheritedSliceTypesChecker::create( cssProperty(), SliceTypes(inheritedImageSlice))); return convertImageSlice(inheritedImageSlice, state.parentStyle()->effectiveZoom()); } InterpolationValue CSSImageSliceInterpolationType::maybeConvertValue( const CSSValue& value, const StyleResolverState&, ConversionCheckers&) const { if (!value.isBorderImageSliceValue()) return nullptr; const CSSBorderImageSliceValue& slice = toCSSBorderImageSliceValue(value); std::unique_ptr<InterpolableList> list = InterpolableList::create(SideIndexCount); const CSSValue* sides[SideIndexCount]; sides[SideTop] = slice.slices().top(); sides[SideRight] = slice.slices().right(); sides[SideBottom] = slice.slices().bottom(); sides[SideLeft] = slice.slices().left(); for (size_t i = 0; i < SideIndexCount; i++) { const CSSPrimitiveValue& side = *toCSSPrimitiveValue(sides[i]); DCHECK(side.isNumber() || side.isPercentage()); list->set(i, InterpolableNumber::create(side.getDoubleValue())); } return InterpolationValue( std::move(list), CSSImageSliceNonInterpolableValue::create(SliceTypes(slice))); } InterpolationValue CSSImageSliceInterpolationType::maybeConvertStandardPropertyUnderlyingValue( const StyleResolverState& state) const { const ComputedStyle& style = *state.style(); return convertImageSlice( ImageSlicePropertyFunctions::getImageSlice(cssProperty(), style), style.effectiveZoom()); } PairwiseInterpolationValue CSSImageSliceInterpolationType::maybeMergeSingles( InterpolationValue&& start, InterpolationValue&& end) const { const SliceTypes& startSliceTypes = toCSSImageSliceNonInterpolableValue(*start.nonInterpolableValue).types(); const SliceTypes& endSliceTypes = toCSSImageSliceNonInterpolableValue(*end.nonInterpolableValue).types(); if (startSliceTypes != endSliceTypes) return nullptr; return PairwiseInterpolationValue(std::move(start.interpolableValue), std::move(end.interpolableValue), std::move(start.nonInterpolableValue)); } void CSSImageSliceInterpolationType::composite( UnderlyingValueOwner& underlyingValueOwner, double underlyingFraction, const InterpolationValue& value, double interpolationFraction) const { const SliceTypes& underlyingTypes = toCSSImageSliceNonInterpolableValue( *underlyingValueOwner.value().nonInterpolableValue) .types(); const SliceTypes& types = toCSSImageSliceNonInterpolableValue(*value.nonInterpolableValue).types(); if (underlyingTypes == types) underlyingValueOwner.mutableValue().interpolableValue->scaleAndAdd( underlyingFraction, *value.interpolableValue); else underlyingValueOwner.set(*this, value); } void CSSImageSliceInterpolationType::applyStandardPropertyValue( const InterpolableValue& interpolableValue, const NonInterpolableValue* nonInterpolableValue, StyleResolverState& state) const { ComputedStyle& style = *state.style(); const InterpolableList& list = toInterpolableList(interpolableValue); const SliceTypes& types = toCSSImageSliceNonInterpolableValue(nonInterpolableValue)->types(); const auto& convertSide = [&types, &list, &style](size_t index) { float value = clampTo<float>(toInterpolableNumber(list.get(index))->value(), 0); return types.isNumber[index] ? Length(value * style.effectiveZoom(), Fixed) : Length(value, Percent); }; LengthBox box(convertSide(SideTop), convertSide(SideRight), convertSide(SideBottom), convertSide(SideLeft)); ImageSlicePropertyFunctions::setImageSlice(cssProperty(), style, ImageSlice(box, types.fill)); } } // namespace blink
3,745
2,350
#pragma once #include "UnrealEnginePython.h" #include "Runtime/Online/Voice/Public/VoiceModule.h" #include "Runtime/Online/Voice/Public/Interfaces/VoiceCapture.h" typedef struct { PyObject_HEAD /* Type-specific fields go here. */ TSharedRef<IVoiceCapture> voice_capture; } ue_PyIVoiceCapture; void ue_python_init_ivoice_capture(PyObject *);
126
2,937
<reponame>fegorsch/ogre /* ----------------------------------------------------------------------------- This source file is part of OGRE (Object-oriented Graphics Rendering Engine) For the latest info, see http://www.ogre3d.org/ Copyright (c) 2000-2014 Torus <NAME> Ltd Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. ----------------------------------------------------------------------------- */ #include "OgreGLVertexArrayObject.h" #include "OgreRoot.h" #include "OgreLogManager.h" #include "OgreGLSLProgramCommon.h" #include "OgreGLRenderSystemCommon.h" namespace Ogre { GLVertexArrayObject::GLVertexArrayObject() : mCreatorContext(0), mVAO(0), mNeedsUpdate(true), mVertexStart(0) { } GLVertexArrayObject::~GLVertexArrayObject() { if(mVAO != 0) { GLRenderSystemCommon* rs = static_cast<GLRenderSystemCommon*>(Root::getSingleton().getRenderSystem()); rs->_destroyVao(mCreatorContext, mVAO); } } void GLVertexArrayObject::bind(GLRenderSystemCommon* rs) { if(mCreatorContext && mCreatorContext != rs->_getCurrentContext()) // VAO is unusable with current context, destroy it { if(mVAO != 0) rs->_destroyVao(mCreatorContext, mVAO); mCreatorContext = 0; mVAO = 0; mNeedsUpdate = true; } if(!mCreatorContext && rs->getCapabilities()->hasCapability(RSC_VAO)) // create VAO lazy or recreate after destruction { mCreatorContext = rs->_getCurrentContext(); mVAO = rs->_createVao(); mNeedsUpdate = true; } rs->_bindVao(mCreatorContext, mVAO); } bool GLVertexArrayObject::needsUpdate(VertexBufferBinding* vertexBufferBinding, size_t vertexStart) { if(mNeedsUpdate) return true; VertexDeclaration::VertexElementList::const_iterator elemIter, elemEnd; elemEnd = mElementList.end(); for (elemIter = mElementList.begin(); elemIter != elemEnd; ++elemIter) { const VertexElement & elem = *elemIter; uint16 source = elem.getSource(); if (!vertexBufferBinding->isBufferBound(source)) continue; // Skip unbound elements VertexElementSemantic sem = elem.getSemantic(); unsigned short elemIndex = elem.getIndex(); uint32 attrib = (uint32)GLSLProgramCommon::getFixedAttributeIndex(sem, elemIndex); const HardwareVertexBufferSharedPtr& vertexBuffer = vertexBufferBinding->getBuffer(source); if (std::find(mAttribsBound.begin(), mAttribsBound.end(), std::make_pair(attrib, vertexBuffer.get())) == mAttribsBound.end()) return true; if (vertexBuffer->isInstanceData() && std::find(mInstanceAttribsBound.begin(), mInstanceAttribsBound.end(), attrib) == mInstanceAttribsBound.end()) return true; } if(vertexStart != mVertexStart) { return true; } return false; } void GLVertexArrayObject::bindToGpu(GLRenderSystemCommon* rs, VertexBufferBinding* vertexBufferBinding, size_t vertexStart) { mAttribsBound.clear(); mInstanceAttribsBound.clear(); VertexDeclaration::VertexElementList::const_iterator elemIter, elemEnd; elemEnd = mElementList.end(); for (elemIter = mElementList.begin(); elemIter != elemEnd; ++elemIter) { const VertexElement& elem = *elemIter; uint16 source = elem.getSource(); if (!vertexBufferBinding->isBufferBound(source)) continue; // Skip unbound elements VertexElementSemantic sem = elem.getSemantic(); unsigned short elemIndex = elem.getIndex(); uint32 attrib = (uint32)GLSLProgramCommon::getFixedAttributeIndex(sem, elemIndex); const HardwareVertexBufferSharedPtr& vertexBuffer = vertexBufferBinding->getBuffer(source); mAttribsBound.push_back(std::make_pair(attrib, vertexBuffer.get())); rs->bindVertexElementToGpu(elem, vertexBuffer, vertexStart); if (vertexBuffer->isInstanceData()) mInstanceAttribsBound.push_back(attrib); } mVertexStart = vertexStart; mNeedsUpdate = false; } }
2,274
1,515
<filename>qigsaw-android/splitinstaller/src/main/java/com/iqiyi/android/qigsaw/core/splitinstall/SplitSessionInstallerImpl.java /* * MIT License * * Copyright (c) 2019-present, iQIYI, Inc. All rights reserved. * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to deal * in the Software without restriction, including without limitation the rights * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell * copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in all * copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ package com.iqiyi.android.qigsaw.core.splitinstall; import com.iqiyi.android.qigsaw.core.splitrequest.splitinfo.SplitInfo; import java.util.List; import java.util.concurrent.Executor; final class SplitSessionInstallerImpl implements SplitSessionInstaller { private final Executor executor; private final SplitInstallSessionManager sessionManager; private final SplitInstaller splitInstaller; SplitSessionInstallerImpl(SplitInstaller splitInstaller, SplitInstallSessionManager sessionManager, Executor executor) { this.splitInstaller = splitInstaller; this.sessionManager = sessionManager; this.executor = executor; } @Override public void install(int sessionId, List<SplitInfo> splitInfoList) { executor.execute(new SplitStartInstallTask(sessionId, splitInstaller, sessionManager, splitInfoList)); } }
617
3,680
// // Copyright 2016 Pixar // // Licensed under the Apache License, Version 2.0 (the "Apache License") // with the following modification; you may not use this file except in // compliance with the Apache License and the following modification to it: // Section 6. Trademarks. is deleted and replaced with: // // 6. Trademarks. This License does not grant permission to use the trade // names, trademarks, service marks, or product names of the Licensor // and its affiliates, except as required to comply with Section 4(c) of // the License and to reproduce the content of the NOTICE file. // // You may obtain a copy of the Apache License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the Apache License with the above modification is // distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY // KIND, either express or implied. See the Apache License for the specific // language governing permissions and limitations under the Apache License. // #include "pxr/usd/usdRender/tokens.h" PXR_NAMESPACE_OPEN_SCOPE UsdRenderTokensType::UsdRenderTokensType() : adjustApertureHeight("adjustApertureHeight", TfToken::Immortal), adjustApertureWidth("adjustApertureWidth", TfToken::Immortal), adjustPixelAspectRatio("adjustPixelAspectRatio", TfToken::Immortal), aspectRatioConformPolicy("aspectRatioConformPolicy", TfToken::Immortal), camera("camera", TfToken::Immortal), color3f("color3f", TfToken::Immortal), cropAperture("cropAperture", TfToken::Immortal), dataType("dataType", TfToken::Immortal), dataWindowNDC("dataWindowNDC", TfToken::Immortal), expandAperture("expandAperture", TfToken::Immortal), full("full", TfToken::Immortal), includedPurposes("includedPurposes", TfToken::Immortal), instantaneousShutter("instantaneousShutter", TfToken::Immortal), intrinsic("intrinsic", TfToken::Immortal), lpe("lpe", TfToken::Immortal), materialBindingPurposes("materialBindingPurposes", TfToken::Immortal), orderedVars("orderedVars", TfToken::Immortal), pixelAspectRatio("pixelAspectRatio", TfToken::Immortal), preview("preview", TfToken::Immortal), primvar("primvar", TfToken::Immortal), productName("productName", TfToken::Immortal), products("products", TfToken::Immortal), productType("productType", TfToken::Immortal), raster("raster", TfToken::Immortal), raw("raw", TfToken::Immortal), renderSettingsPrimPath("renderSettingsPrimPath", TfToken::Immortal), resolution("resolution", TfToken::Immortal), sourceName("sourceName", TfToken::Immortal), sourceType("sourceType", TfToken::Immortal), allTokens({ adjustApertureHeight, adjustApertureWidth, adjustPixelAspectRatio, aspectRatioConformPolicy, camera, color3f, cropAperture, dataType, dataWindowNDC, expandAperture, full, includedPurposes, instantaneousShutter, intrinsic, lpe, materialBindingPurposes, orderedVars, pixelAspectRatio, preview, primvar, productName, products, productType, raster, raw, renderSettingsPrimPath, resolution, sourceName, sourceType }) { } TfStaticData<UsdRenderTokensType> UsdRenderTokens; PXR_NAMESPACE_CLOSE_SCOPE
1,252
314
package com.lynnchurch.pulltorefresh; import android.content.Context; import android.support.v7.widget.RecyclerView; import android.view.LayoutInflater; import android.view.View; import android.view.ViewGroup; import android.widget.TextView; import com.jingchen.pulltorefresh.WrapRecyclerView; import com.lsjwzh.widget.materialloadingprogressbar.CircleProgressBar; import java.util.ArrayList; /** * 自动加载更多适配器 */ public abstract class BaseAutoLoadMoreAdapter<T> extends RecyclerView.Adapter<RecyclerView.ViewHolder> { private static final String TAG = BaseAutoLoadMoreAdapter.class.getSimpleName(); protected Context mContext; protected ArrayList<T> mData; private OnItemClickListener mOnItemClickListener; private OnLoadmoreListener mOnLoadmoreListener; private int mLastPosition; // 正常项最后一项的位置 private WrapRecyclerView mParentView; private int mParentHeight; // item父视图的高度 private int mItemsHeight; // 所有item的总高度 private TextView tv_hint; private CircleProgressBar progressBar; public BaseAutoLoadMoreAdapter(Context context, WrapRecyclerView recyclerView, ArrayList<T> data) { mContext = context; mParentView = recyclerView; initFooter(recyclerView); mData = data; mLastPosition = mData.size() - 1; } private void initFooter(WrapRecyclerView recyclerView) { View footer = LayoutInflater.from(mContext).inflate(R.layout.loadmore, null); tv_hint = (TextView) footer.findViewById(R.id.tv_hint); progressBar = (CircleProgressBar) footer.findViewById(R.id.progressBar); recyclerView.addFootView(footer); showLoading(false); } @Override final public RecyclerView.ViewHolder onCreateViewHolder(ViewGroup parent, int viewType) { mParentHeight = parent.getHeight(); return onCreateBaseViewHolder(parent, viewType); } public abstract BaseViewHolder onCreateBaseViewHolder(ViewGroup parent, int viewType); public abstract int getItemHeight(RecyclerView.ViewHolder holder); public void resetItemHeight() { mItemsHeight = 0; } @Override final public void onBindViewHolder(RecyclerView.ViewHolder holder, final int position) { if(!isOverParent()) { showLoading(false); mItemsHeight+=getItemHeight(holder); } if (holder instanceof BaseViewHolder) { final BaseViewHolder viewHolder = (BaseViewHolder) holder; if (null != mOnItemClickListener) { viewHolder.itemView.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { mOnItemClickListener.onItemClick(viewHolder.itemView, position); } }); viewHolder.itemView.setOnLongClickListener(new View.OnLongClickListener() { @Override public boolean onLongClick(View v) { mOnItemClickListener.onItemLongClick(viewHolder.itemView, position); return true; } }); } onBindBaseViewHolder(viewHolder, position); } if (position + 1 > mLastPosition && null != mOnLoadmoreListener && isOverParent()) { showLoading(true); mOnLoadmoreListener.onLoadmore(); mLastPosition = mData.size(); } } public abstract void onBindBaseViewHolder(BaseViewHolder holder, final int position); /** * item是否撑满父视图 * * @return */ public boolean isOverParent() { return mParentHeight < mItemsHeight + mParentView.getHeaderHeight() ? true : false; } @Override public int getItemCount() { return mData.size(); } public static class BaseViewHolder extends RecyclerView.ViewHolder { View itemView; public BaseViewHolder(View v) { super(v); itemView = v; } } /** * 当加载失败 */ public void onFailed() { progressBar.setVisibility(View.GONE); tv_hint.setVisibility(View.VISIBLE); tv_hint.setText("点击重试"); tv_hint.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { if (null != mOnLoadmoreListener) { progressBar.setVisibility(View.VISIBLE); tv_hint.setVisibility(View.GONE); mOnLoadmoreListener.onLoadmore(); } } }); } /** * 当没有更多 */ public void onNothing() { progressBar.setVisibility(View.GONE); tv_hint.setVisibility(View.VISIBLE); tv_hint.setText("没有更多"); tv_hint.setOnClickListener(null); } /** * 显示加载视图 * * @param show */ private void showLoading(boolean show) { if (show) { progressBar.setVisibility(View.VISIBLE); } else { progressBar.setVisibility(View.GONE); } } /** * 当数据改变时调用,替换notifyDataSetChanged() */ public void notifyDataChanged() { notifyDataSetChanged(); mParentView.requestLayout(); } /** * 设置Item监听器 * * @param onItemClickListener */ public void setOnItemClickListener(OnItemClickListener onItemClickListener) { mOnItemClickListener = onItemClickListener; } /** * Item点击监听 */ public interface OnItemClickListener { void onItemClick(View view, int position); void onItemLongClick(View view, int position); } /** * 设置加载更多监听 * * @param listener */ public void setOnLoadmoreListener(OnLoadmoreListener listener) { mOnLoadmoreListener = listener; } /** * 加载更多监听 */ public interface OnLoadmoreListener { void onLoadmore(); } }
3,030
1,241
import math import os from time import time from typing import List, Dict, Union, Tuple from tensorflow.python.keras.models import load_model from donkeycar.config import Config from donkeycar.parts.keras import KerasPilot from donkeycar.parts.interpreter import keras_model_to_tflite, \ saved_model_to_tensor_rt from donkeycar.pipeline.database import PilotDatabase from donkeycar.pipeline.sequence import TubRecord, TubSequence, TfmIterator from donkeycar.pipeline.types import TubDataset from donkeycar.pipeline.augmentations import ImageAugmentation from donkeycar.utils import get_model_by_type, normalize_image, train_test_split import tensorflow as tf import numpy as np class BatchSequence(object): """ The idea is to have a shallow sequence with types that can hydrate themselves to np.ndarray initially and later into the types required by tf.data (i.e. dictionaries or np.ndarrays). """ def __init__(self, model: KerasPilot, config: Config, records: List[TubRecord], is_train: bool) -> None: self.model = model self.config = config self.sequence = TubSequence(records) self.batch_size = self.config.BATCH_SIZE self.is_train = is_train self.augmentation = ImageAugmentation(config, 'AUGMENTATIONS') self.transformation = ImageAugmentation(config, 'TRANSFORMATIONS') self.pipeline = self._create_pipeline() def __len__(self) -> int: return math.ceil(len(self.pipeline) / self.batch_size) def image_processor(self, img_arr): """ Transformes the images and augments if in training. Then normalizes it. """ img_arr = self.transformation.run(img_arr) if self.is_train: img_arr = self.augmentation.run(img_arr) norm_img = normalize_image(img_arr) return norm_img def _create_pipeline(self) -> TfmIterator: """ This can be overridden if more complicated pipelines are required """ # 1. Initialise TubRecord -> x, y transformations def get_x(record: TubRecord) -> Dict[str, Union[float, np.ndarray]]: """ Extracting x from record for training""" out_tuple = self.model.x_transform_and_process( record, self.image_processor) # convert tuple to dictionary which is understood by tf.data out_dict = self.model.x_translate(out_tuple) return out_dict def get_y(record: TubRecord) -> Dict[str, Union[float, np.ndarray]]: """ Extracting y from record for training """ y0 = self.model.y_transform(record) y1 = self.model.y_translate(y0) return y1 # 2. Build pipeline using the transformations pipeline = self.sequence.build_pipeline(x_transform=get_x, y_transform=get_y) return pipeline def create_tf_data(self) -> tf.data.Dataset: """ Assembles the tf data pipeline """ dataset = tf.data.Dataset.from_generator( generator=lambda: self.pipeline, output_types=self.model.output_types(), output_shapes=self.model.output_shapes()) return dataset.repeat().batch(self.batch_size) def get_model_train_details(database: PilotDatabase, model: str = None) \ -> Tuple[str, int]: if not model: model_name, model_num = database.generate_model_name() else: model_name, model_num = os.path.abspath(model), 0 return model_name, model_num def train(cfg: Config, tub_paths: str, model: str = None, model_type: str = None, transfer: str = None, comment: str = None) \ -> tf.keras.callbacks.History: """ Train the model """ database = PilotDatabase(cfg) if model_type is None: model_type = cfg.DEFAULT_MODEL_TYPE model_path, model_num = \ get_model_train_details(database, model) base_path = os.path.splitext(model_path)[0] kl = get_model_by_type(model_type, cfg) if transfer: kl.load(transfer) if cfg.PRINT_MODEL_SUMMARY: print(kl.interpreter.model.summary()) tubs = tub_paths.split(',') all_tub_paths = [os.path.expanduser(tub) for tub in tubs] dataset = TubDataset(config=cfg, tub_paths=all_tub_paths, seq_size=kl.seq_size()) training_records, validation_records \ = train_test_split(dataset.get_records(), shuffle=True, test_size=(1. - cfg.TRAIN_TEST_SPLIT)) print(f'Records # Training {len(training_records)}') print(f'Records # Validation {len(validation_records)}') # We need augmentation in validation when using crop / trapeze training_pipe = BatchSequence(kl, cfg, training_records, is_train=True) validation_pipe = BatchSequence(kl, cfg, validation_records, is_train=False) tune = tf.data.experimental.AUTOTUNE dataset_train = training_pipe.create_tf_data().prefetch(tune) dataset_validate = validation_pipe.create_tf_data().prefetch(tune) train_size = len(training_pipe) val_size = len(validation_pipe) assert val_size > 0, "Not enough validation data, decrease the batch " \ "size or add more data." history = kl.train(model_path=model_path, train_data=dataset_train, train_steps=train_size, batch_size=cfg.BATCH_SIZE, validation_data=dataset_validate, validation_steps=val_size, epochs=cfg.MAX_EPOCHS, verbose=cfg.VERBOSE_TRAIN, min_delta=cfg.MIN_DELTA, patience=cfg.EARLY_STOP_PATIENCE, show_plot=cfg.SHOW_PLOT) if getattr(cfg, 'CREATE_TF_LITE', True): tf_lite_model_path = f'{base_path}.tflite' keras_model_to_tflite(model_path, tf_lite_model_path) if getattr(cfg, 'CREATE_TENSOR_RT', False): # load h5 (ie. keras) model model_rt = load_model(model_path) # save in tensorflow savedmodel format (i.e. directory) model_rt.save(f'{base_path}.savedmodel') # pass savedmodel to the rt converter saved_model_to_tensor_rt(f'{base_path}.savedmodel', f'{base_path}.trt') database_entry = { 'Number': model_num, 'Name': os.path.basename(base_path), 'Type': str(kl), 'Tubs': tub_paths, 'Time': time(), 'History': history.history, 'Transfer': os.path.basename(transfer) if transfer else None, 'Comment': comment, 'Config': str(cfg) } database.add_entry(database_entry) database.write() return history
3,030
435
<filename>pycon-tw-2021/videos/supercharging-your-jupyter-and-ml-workflow-with-primehub-chia-liang-kao-pycon-taiwan-2021.json<gh_stars>100-1000 { "description": "Title: Supercharging your Jupyter and Machine Learning Workflow with PrimeHub \u2013 <NAME> \n\nDay 2, 10:35-11:20\n\nAbstract\n\nJupyter has become a critical component of the machine learning life cycle. However scaled enterprise deployments and making the Data Science Experience frictionless remain challenging. We address a few common issues with PrimeHub, an open-source enterprise offering based on JupyterHub, with the capability to work with popular ML tooling like mlflow, labelstudio, and streamlit. This talk also investigates the MLOps trends adjacent to the Jupyter ecosystem.\n\nDescription\n\nThis talk is intended for audience interested in larger scale Jupyter environment deployment in their organisation, particularly for machine learning applications.\n\nPrimeHub is an open-source enterprise offering based on JupyterHub, which also allows orchestrating tools like mlflow, labelstudio, and streamlit, to assemble your own end-to-end ML toolbox for your team. We address a few common hurdles:\n\nadvanced resource scheduling for heterogenous GPU clusters\nmulti-tenancy and project isolations\ncustomizable and consistent ML environments per project\nmanaging jupyterlab extensions for teams\nhybrid development environment such as vscode and ssh within notebook instances\nauthorization and data access management\nusage reporting: allocation and utilization\nWe also investigate a few trends adjacent to the day-to-day jupyter environment used by data scientists and data engineers, where the roles become more cross functional in the age of MLOps:\n\nWorking with job schedulers from within Notebook\nManaging ML model deployments, CI/CD integration\nModel monitoring and retraining\n\nSlides not uploaded by the speaker.\nHackMD: https://hackmd.io/@pycontw/2021/%2F%40pycontw%2Frk-phz9zF\n\nSpeaker: <NAME>\n\nclkao (<NAME>) has been an open source software developer since 2000. He believes that good collaboration model and tools drive innovation. In 2013, he created SVK, a distributed version control system that helps developers collaborate. He co-founded the g0v.tw community in 2012, advocating information transparency and digital-activism through open source model. g0v.tw was awarded as \"Digital Communities: Award of Distinction\" by Prix Ars Electronica 2018. He started InfuseAI in 2018 to enable data scientists to thrive, and to help wider adoption of AI across industries.", "speakers": [ "<NAME>" ], "recorded": "2021-10-03", "title": "Supercharging your Jupyter and ML Workflow with PrimeHub \u2013 Chia-liang Kao (PyCon Taiwan 2021)", "thumbnail_url": "https://i.ytimg.com/vi/qf0U1JIgCFY/hqdefault.jpg", "videos": [ { "type": "youtube", "url": "https://www.youtube.com/watch?v=qf0U1JIgCFY" } ] }
869
892
{ "schema_version": "1.2.0", "id": "GHSA-3vv6-22qm-hgwq", "modified": "2022-05-13T01:36:56Z", "published": "2022-05-13T01:36:56Z", "aliases": [ "CVE-2017-2348" ], "details": "The Juniper Enhanced jdhcpd daemon may experience high CPU utilization, or crash and restart upon receipt of an invalid IPv6 UDP packet. Both high CPU utilization and repeated crashes of the jdhcpd daemon can result in a denial of service as DHCP service is interrupted. No other Juniper Networks products or platforms are affected by this issue. Affected releases are Juniper Networks Junos OS 14.1X53 prior to 14.1X53-D12, 14.1X53-D38, 14.1X53-D40 on QFX, EX, QFabric System; 15.1 prior to 15.1F2-S18, 15.1R4 on all products and platforms; 15.1X49 prior to 15.1X49-D80 on SRX; 15.1X53 prior to 15.1X53-D51, 15.1X53-D60 on NFX, QFX, EX.", "severity": [ { "type": "CVSS_V3", "score": "CVSS:3.0/AV:N/AC:L/PR:N/UI:N/S:U/C:N/I:N/A:H" } ], "affected": [ ], "references": [ { "type": "ADVISORY", "url": "https://nvd.nist.gov/vuln/detail/CVE-2017-2348" }, { "type": "WEB", "url": "https://kb.juniper.net/JSA10800" }, { "type": "WEB", "url": "http://www.securitytracker.com/id/1038899" } ], "database_specific": { "cwe_ids": [ "CWE-400" ], "severity": "HIGH", "github_reviewed": false } }
627
325
from .process import Processing, Mask, F from .partition import Partition, TimePartition
21
672
<filename>xnu-4903.241.1/bsd/arm/disklabel.h /* * Copyright (c) 2000-2007 Apple Inc. All rights reserved. */ #ifndef _MACHINE_DISKLABEL_H_ #define _MACHINE_DISKLABEL_H_ #include <sys/appleapiopts.h> #ifdef __APPLE_API_OBSOLETE #define LABELSECTOR (1024 / DEV_BSIZE) /* sector containing label */ #define LABELOFFSET 0 /* offset of label in sector */ #define MAXPARTITIONS 8 /* number of partitions */ #define RAW_PART 2 /* raw partition: xx?c */ /* Just a dummy */ struct cpu_disklabel { int cd_dummy; /* must have one element. */ }; #endif /* __APPLE_API_OBSOLETE */ #endif /* _MACHINE_DISKLABEL_H_ */
256
398
<gh_stars>100-1000 #!/usr/bin/env python3 import argparse, csv, sys, pickle, collections, time from common import * if len(sys.argv) == 1: sys.argv.append('-h') parser = argparse.ArgumentParser() parser.add_argument('tr_src_path', type=str) parser.add_argument('va_src_path', type=str) parser.add_argument('tr_app_dst_path', type=str) parser.add_argument('va_app_dst_path', type=str) parser.add_argument('tr_site_dst_path', type=str) parser.add_argument('va_site_dst_path', type=str) args = vars(parser.parse_args()) FIELDS = ['id','click','hour','banner_pos','device_id','device_ip','device_model','device_conn_type','C14','C17','C20','C21'] NEW_FIELDS = FIELDS+['pub_id','pub_domain','pub_category','device_id_count','device_ip_count','user_count','smooth_user_hour_count','user_click_histroy'] id_cnt = collections.defaultdict(int) ip_cnt = collections.defaultdict(int) user_cnt = collections.defaultdict(int) user_hour_cnt = collections.defaultdict(int) start = time.time() def scan(path): for i, row in enumerate(csv.DictReader(open(path)), start=1): if i % 1000000 == 0: sys.stderr.write('{0:6.0f} {1}m\n'.format(time.time()-start,int(i/1000000))) user = def_user(row) id_cnt[row['device_id']] += 1 ip_cnt[row['device_ip']] += 1 user_cnt[user] += 1 user_hour_cnt[user+'-'+row['hour']] += 1 history = collections.defaultdict(lambda: {'history': '', 'buffer': '', 'prev_hour': ''}) def gen_data(src_path, dst_app_path, dst_site_path, is_train): reader = csv.DictReader(open(src_path)) writer_app = csv.DictWriter(open(dst_app_path, 'w'), NEW_FIELDS) writer_site = csv.DictWriter(open(dst_site_path, 'w'), NEW_FIELDS) writer_app.writeheader() writer_site.writeheader() for i, row in enumerate(reader, start=1): if i % 1000000 == 0: sys.stderr.write('{0:6.0f} {1}m\n'.format(time.time()-start,int(i/1000000))) new_row = {} for field in FIELDS: new_row[field] = row[field] new_row['device_id_count'] = id_cnt[row['device_id']] new_row['device_ip_count'] = ip_cnt[row['device_ip']] user, hour = def_user(row), row['hour'] new_row['user_count'] = user_cnt[user] new_row['smooth_user_hour_count'] = str(user_hour_cnt[user+'-'+hour]) if has_id_info(row): if history[user]['prev_hour'] != row['hour']: history[user]['history'] = (history[user]['history'] + history[user]['buffer'])[-4:] history[user]['buffer'] = '' history[user]['prev_hour'] = row['hour'] new_row['user_click_histroy'] = history[user]['history'] if is_train: history[user]['buffer'] += row['click'] else: new_row['user_click_histroy'] = '' if is_app(row): new_row['pub_id'] = row['app_id'] new_row['pub_domain'] = row['app_domain'] new_row['pub_category'] = row['app_category'] writer_app.writerow(new_row) else: new_row['pub_id'] = row['site_id'] new_row['pub_domain'] = row['site_domain'] new_row['pub_category'] = row['site_category'] writer_site.writerow(new_row) scan(args['tr_src_path']) scan(args['va_src_path']) print('======================scan complete======================') gen_data(args['tr_src_path'], args['tr_app_dst_path'], args['tr_site_dst_path'], True) gen_data(args['va_src_path'], args['va_app_dst_path'], args['va_site_dst_path'], False)
1,643
2,389
/* Copyright 2020 The Kubernetes Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ package io.kubernetes.client.informer.cache; import java.util.List; public interface Store<ApiType> { /** * add inserts an item into the store. * * @param obj specific obj */ void add(ApiType obj); /** * update sets an item in the store to its updated state. * * @param obj specific obj */ void update(ApiType obj); /** * delete removes an item from the store. * * @param obj specific obj */ void delete(ApiType obj); /** * Replace will delete the contents of 'c', using instead the given list. * * @param list list of objects * @param resourceVersion specific resource version */ void replace(List<ApiType> list, String resourceVersion); /** resync will send a resync event for each item. */ void resync(); /** * listKeys returns a list of all the keys of the object currently in the store. * * @return list of all keys */ List<String> listKeys(); /** * get returns the requested item. * * @param obj specific obj * @return the requested item if exist */ Object get(ApiType obj); /** * getByKey returns the request item with specific key. * * @param key specific key * @return the request item */ ApiType getByKey(String key); /** * list returns a list of all the items. * * @return list of all the items */ List<ApiType> list(); }
603
711
<reponame>jingetiema2100/MicroCommunity<filename>service-store/src/main/java/com/java110/store/bmo/contractType/IDeleteContractTypeBMO.java<gh_stars>100-1000 package com.java110.store.bmo.contractType; import com.java110.po.contractType.ContractTypePo; import org.springframework.http.ResponseEntity; public interface IDeleteContractTypeBMO { /** * 修改合同类型 * add by wuxw * @param contractTypePo * @return */ ResponseEntity<String> delete(ContractTypePo contractTypePo); }
198
575
<reponame>zealoussnow/chromium<gh_stars>100-1000 // Copyright 2018 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "chrome/elevation_service/elevated_recovery_impl.h" #include <objbase.h> #include <string> #include <utility> #include "base/base_paths.h" #include "base/bind.h" #include "base/callback_helpers.h" #include "base/command_line.h" #include "base/files/file_enumerator.h" #include "base/files/file_path.h" #include "base/files/file_util.h" #include "base/files/scoped_temp_dir.h" #include "base/logging.h" #include "base/path_service.h" #include "base/process/launch.h" #include "base/process/process.h" #include "base/strings/utf_string_conversions.h" #include "base/version.h" #include "base/win/scoped_process_information.h" #include "chrome/install_static/install_util.h" #include "third_party/zlib/google/zip.h" namespace elevation_service { namespace { // Input CRX files over 10 MB are considered invalid. constexpr int64_t kMaxFileSize = 10u * 1000u * 1000u; // The hard-coded file name that the Recovery CRX is copied to. constexpr base::FilePath::CharType kCRXFileName[] = FILE_PATH_LITERAL("ChromeRecoveryCRX.crx"); // The hard-coded Recovery subdirectory where the CRX is unpacked and executed. constexpr base::FilePath::CharType kRecoveryDirectory[] = FILE_PATH_LITERAL("ChromeRecovery"); // The hard-coded Recovery executable. This file comes from the CRX, and we // create an elevated process from it. constexpr base::FilePath::CharType kRecoveryExeName[] = FILE_PATH_LITERAL("ChromeRecovery.exe"); // The hard-coded SHA256 of the SubjectPublicKeyInfo used to sign the Recovery // CRX which contains ChromeRecovery.exe. std::vector<uint8_t> GetRecoveryCRXHash() { return std::vector<uint8_t>{0x5f, 0x94, 0xe0, 0x3c, 0x64, 0x30, 0x9f, 0xbc, 0xfe, 0x00, 0x9a, 0x27, 0x3e, 0x52, 0xbf, 0xa5, 0x84, 0xb9, 0xb3, 0x75, 0x07, 0x29, 0xde, 0xfa, 0x32, 0x76, 0xd9, 0x93, 0xb5, 0xa3, 0xce, 0x02}; } // This function is only meant to be called when a Windows API function errors // out, and the corresponding ::GetLastResult is expected to be set to an error. // There could be cases where ::GetLastError() is not set correctly, this // function returns E_FAIL in those cases. HRESULT HRESULTFromLastError() { const auto error_code = ::GetLastError(); return (error_code != ERROR_SUCCESS) ? HRESULT_FROM_WIN32(error_code) : E_FAIL; } // Opens and returns the COM caller's |process| given the process id, or the // current process if |proc_id| is 0. HRESULT OpenCallingProcess(uint32_t proc_id, base::Process* process) { DCHECK(proc_id); DCHECK(process); HRESULT hr = ::CoImpersonateClient(); if (FAILED(hr)) return hr; base::ScopedClosureRunner revert_to_self( base::BindOnce([]() { ::CoRevertToSelf(); })); *process = base::Process::OpenWithAccess(proc_id, PROCESS_DUP_HANDLE); return process->IsValid() ? S_OK : HRESULTFromLastError(); } // Opens and returns a base::File instance for the |file_path|. We impersonate // the COM caller when opening the base::File instance. This is to ensure that // the COM caller has access to the file. HRESULT OpenFileImpersonated(const base::FilePath& file_path, int flags, base::File* file) { DCHECK(file); HRESULT hr = ::CoImpersonateClient(); if (FAILED(hr)) return hr; base::ScopedClosureRunner revert_to_self( base::BindOnce([]() { ::CoRevertToSelf(); })); file->Initialize(file_path, flags); if (!file->IsValid()) return HRESULTFromLastError(); if (::GetFileType(file->GetPlatformFile()) != FILE_TYPE_DISK) return E_INVALIDARG; int64_t from_length = file->GetLength(); return from_length > 0 && from_length < kMaxFileSize ? S_OK : E_INVALIDARG; } // Opens |from| while impersonating the COM caller, and then copies the contents // of |from| to |to|. |from| is opened impersonated to ensure that the COM // caller has access to the file. HRESULT CopyFileImpersonated(const base::FilePath from, const base::FilePath& to) { base::File from_file; HRESULT hr = OpenFileImpersonated(from, base::File::FLAG_READ | base::File::FLAG_OPEN | base::File::FLAG_SEQUENTIAL_SCAN, &from_file); if (FAILED(hr)) return hr; base::File to_file; to_file.Initialize(to, base::File::FLAG_WRITE | base::File::FLAG_CREATE_ALWAYS | base::File::FLAG_SEQUENTIAL_SCAN); if (!to_file.IsValid()) return HRESULTFromLastError(); constexpr size_t kBufferSize = 0x10000; std::vector<char> buffer(kBufferSize); for (uint64_t total_bytes_read = 0;;) { const int bytes_read = from_file.ReadAtCurrentPos(buffer.data(), buffer.size()); if (bytes_read < 0) return HRESULTFromLastError(); if (bytes_read == 0) return S_OK; total_bytes_read += bytes_read; if (total_bytes_read > kMaxFileSize) return E_INVALIDARG; const int bytes_written = to_file.WriteAtCurrentPos(&buffer[0], bytes_read); if (bytes_written < 0) return HRESULTFromLastError(); if (bytes_written != bytes_read) return E_UNEXPECTED; } NOTREACHED(); return S_OK; } // Validates the provided CRX using the |crx_hash|, and if validation succeeds, // unpacks the CRX under |unpack_under_path|. Returns the unpacked CRX // directory in |unpacked_crx_dir|. HRESULT ValidateAndUnpackCRX(const base::FilePath& from_crx_path, const crx_file::VerifierFormat& crx_format, const std::vector<uint8_t>& crx_hash, const base::FilePath& unpack_under_path, base::ScopedTempDir* unpacked_crx_dir) { DCHECK(unpacked_crx_dir); base::ScopedTempDir to_dir; if (!to_dir.CreateUniqueTempDirUnderPath(unpack_under_path)) return HRESULTFromLastError(); const base::FilePath to_crx_path = to_dir.GetPath().Append(kCRXFileName); // Copy |from_crx_path| impersonated. This is to prevent us from copying files // that may not be accessible to the calling COM user. HRESULT hr = CopyFileImpersonated(from_crx_path, to_crx_path); if (FAILED(hr)) return hr; std::string public_key; if (crx_file::Verify(to_crx_path, crx_format, {crx_hash}, {}, &public_key, nullptr, /*compressed_verified_contents=*/nullptr) != crx_file::VerifierResult::OK_FULL) { return CRYPT_E_NO_MATCH; } if (!zip::Unzip(to_crx_path, to_dir.GetPath())) return E_UNEXPECTED; LOG_IF(WARNING, !base::DeleteFile(to_crx_path)); LOG_IF(WARNING, !unpacked_crx_dir->Set(to_dir.Take())); return S_OK; } // Runs the executable |path_and_name| with the provided |args|. The returned // |proc_handle| is a process handle that is valid for the |calling_process| // process. HRESULT LaunchCmd(const base::CommandLine& command_line, const base::Process& calling_process, base::win::ScopedHandle* proc_handle) { DCHECK(!command_line.GetCommandLineString().empty()); DCHECK(proc_handle); base::LaunchOptions options = {}; options.feedback_cursor_off = true; base::GetTempDir(&options.current_directory); base::Process proc = base::LaunchProcess(command_line, options); if (!proc.IsValid()) return HRESULTFromLastError(); HANDLE duplicate_proc_handle = nullptr; constexpr DWORD desired_access = PROCESS_QUERY_LIMITED_INFORMATION | SYNCHRONIZE; bool res = ::DuplicateHandle( ::GetCurrentProcess(), // Current process. proc.Handle(), // Process handle to duplicate. calling_process.Handle(), // Process receiving the handle. &duplicate_proc_handle, // Duplicated handle. desired_access, // Access requested for the new handle. FALSE, // Don't inherit the new handle. 0) != 0; if (!res) return HRESULTFromLastError(); proc_handle->Set(duplicate_proc_handle); return S_OK; } HRESULT ValidateCRXArgs(const std::wstring& browser_appid, const std::wstring& browser_version, const std::wstring& session_id) { if (!browser_appid.empty()) { GUID guid = {}; HRESULT hr = ::IIDFromString(browser_appid.c_str(), &guid); if (FAILED(hr)) return hr; } const base::Version version(base::WideToASCII(browser_version)); if (!version.IsValid()) return E_INVALIDARG; GUID session_guid = {}; return ::IIDFromString(session_id.c_str(), &session_guid); } // Deletes all the files and subdirectories within |directory_path|. Errors are // ignored. void DeleteDirectoryFiles(const base::FilePath& directory_path) { base::FileEnumerator file_enum( directory_path, false, base::FileEnumerator::FILES | base::FileEnumerator::DIRECTORIES); for (base::FilePath current = file_enum.Next(); !current.empty(); current = file_enum.Next()) { base::DeletePathRecursively(current); } } // Schedules deletion after reboot of |dir_name| as well as all the files and // subdirectories within |dir_name|. Errors are ignored. void ScheduleDirectoryForDeletion(const base::FilePath& dir_name) { // First schedule all the files within |dir_name| for deletion. base::FileEnumerator file_enum(dir_name, false, base::FileEnumerator::FILES); for (base::FilePath file = file_enum.Next(); !file.empty(); file = file_enum.Next()) { base::DeleteFileAfterReboot(file); } // Then recurse to all the subdirectories. base::FileEnumerator dir_enum(dir_name, false, base::FileEnumerator::DIRECTORIES); for (base::FilePath sub_dir = dir_enum.Next(); !sub_dir.empty(); sub_dir = dir_enum.Next()) { ScheduleDirectoryForDeletion(sub_dir); } // Now schedule the empty directory itself. base::DeleteFileAfterReboot(dir_name); } // Returns the ChromeRecovery directory under Google\Chrome. For machine // installs, this directory is under %ProgramFiles%, which is writeable only by // adminstrators. We use this secure directory to validate and unpack the CRX to // prevent tampering. HRESULT GetChromeRecoveryDirectory(base::FilePath* dir) { base::FilePath recovery_dir; if (!base::PathService::Get(base::DIR_EXE, &recovery_dir)) return HRESULTFromLastError(); recovery_dir = recovery_dir.DirName().DirName().Append(kRecoveryDirectory); *dir = std::move(recovery_dir); return S_OK; } } // namespace HRESULT CleanupChromeRecoveryDirectory() { base::FilePath recovery_dir; HRESULT hr = GetChromeRecoveryDirectory(&recovery_dir); if (FAILED(hr)) return hr; DeleteDirectoryFiles(recovery_dir); return S_OK; } HRESULT RunChromeRecoveryCRX(const base::FilePath& crx_path, const std::wstring& browser_appid, const std::wstring& browser_version, const std::wstring& session_id, uint32_t caller_proc_id, base::win::ScopedHandle* proc_handle) { if (crx_path.empty() || !caller_proc_id || !proc_handle) return E_INVALIDARG; HRESULT hr = ValidateCRXArgs(browser_appid, browser_version, session_id); if (FAILED(hr)) return hr; base::CommandLine args(base::CommandLine::NO_PROGRAM); if (!browser_appid.empty()) args.AppendSwitchNative("appguid", browser_appid); args.AppendSwitchNative("browser-version", browser_version); args.AppendSwitchNative("sessionid", session_id); args.AppendSwitch("system"); base::FilePath unpack_dir; hr = GetChromeRecoveryDirectory(&unpack_dir); if (FAILED(hr)) return hr; return RunCRX(crx_path, args, crx_file::VerifierFormat::CRX3_WITH_PUBLISHER_PROOF, GetRecoveryCRXHash(), unpack_dir, base::FilePath(kRecoveryExeName), caller_proc_id, proc_handle); } HRESULT RunCRX(const base::FilePath& crx_path, const base::CommandLine& args, const crx_file::VerifierFormat& crx_format, const std::vector<uint8_t>& crx_hash, const base::FilePath& unpack_under_path, const base::FilePath& exe_filename, uint32_t caller_proc_id, base::win::ScopedHandle* proc_handle) { DCHECK(!crx_path.empty()); DCHECK(!crx_hash.empty()); DCHECK(!unpack_under_path.empty()); DCHECK(!exe_filename.empty()); DCHECK(caller_proc_id); DCHECK(proc_handle); base::Process calling_process; HRESULT hr = OpenCallingProcess(caller_proc_id, &calling_process); if (FAILED(hr)) return hr; base::ScopedTempDir unpacked_crx_dir; hr = ValidateAndUnpackCRX(crx_path, crx_format, crx_hash, unpack_under_path, &unpacked_crx_dir); if (FAILED(hr)) return hr; const base::FilePath path_and_name = unpacked_crx_dir.GetPath().Append(exe_filename); base::CommandLine command_line(path_and_name); command_line.AppendArguments(args, false); base::win::ScopedHandle scoped_proc_handle; hr = LaunchCmd(command_line, calling_process, &scoped_proc_handle); if (FAILED(hr)) return hr; ScheduleDirectoryForDeletion(unpacked_crx_dir.Take()); *proc_handle = std::move(scoped_proc_handle); return hr; } } // namespace elevation_service
5,586
357
# Copyright 2019 The Blueqat Developers # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from collections import Counter import pytest from blueqat.utils import to_inttuple @pytest.mark.parametrize('arg, expect', [ ("01011", (0, 1, 0, 1, 1)), ({"00011": 2, "10100": 3}, {(0, 0, 0, 1, 1): 2, (1, 0, 1, 0, 0): 3}), (Counter({"00011": 2, "10100": 3}), Counter({(0, 0, 0, 1, 1): 2, (1, 0, 1, 0, 0): 3})) ]) def test_to_inttuple(arg, expect): assert to_inttuple(arg) == expect
336
1,179
// SPDX-License-Identifier: BSD-2-Clause /* * Copyright (c) 2014, STMicroelectronics International N.V. */ #include <compiler.h> #include <link.h> #include <stdbool.h> #include <stdlib.h> #include <string.h> #include <sys/queue.h> #include <tee_api.h> #include <tee_ta_api.h> #include <tee_internal_api_extensions.h> #include <user_ta_header.h> #include <utee_syscalls.h> #include <tee_arith_internal.h> #include <malloc.h> #include "tee_api_private.h" struct ta_session { uint32_t session_id; void *session_ctx; TAILQ_ENTRY(ta_session) link; }; static TAILQ_HEAD(ta_sessions, ta_session) ta_sessions = TAILQ_HEAD_INITIALIZER(ta_sessions); static bool init_done; /* From user_ta_header.c, built within TA */ extern uint8_t ta_heap[]; extern const size_t ta_heap_size; extern struct ta_head ta_head; uint32_t ta_param_types; TEE_Param ta_params[TEE_NUM_PARAMS]; struct __elf_phdr_info __elf_phdr_info; struct phdr_info { struct dl_phdr_info info; TAILQ_ENTRY(phdr_info) link; }; static TAILQ_HEAD(phdr_info_head, phdr_info) __phdr_info_head = TAILQ_HEAD_INITIALIZER(__phdr_info_head); /* * Keep track of how many modules have been initialized so that subsequent * dlopen() calls will not run the same initializers again */ static size_t _num_mod_init; static int _init_iterate_phdr_cb(struct dl_phdr_info *info, size_t size __unused, void *data) { struct phdr_info *qe = NULL; size_t *count = data; qe = malloc(sizeof(*qe)); if (!qe) { EMSG("init/fini: out of memory"); abort(); } qe->info = *info; TAILQ_INSERT_TAIL(&__phdr_info_head, qe, link); (*count)++; return 0; } static void _get_fn_array(struct dl_phdr_info *info, Elf_Sword tag_a, Elf_Sword tag_s, void (***fn)(void), size_t *num_fn) { const Elf_Phdr *phdr = NULL; Elf_Dyn *dyn = NULL; size_t num_dyn = 0; size_t i = 0; size_t j = 0; for (i = 0; i < info->dlpi_phnum; i++) { phdr = info->dlpi_phdr + i; if (phdr->p_type != PT_DYNAMIC) continue; num_dyn = phdr->p_memsz / sizeof(Elf_Dyn); dyn = (Elf_Dyn *)(phdr->p_vaddr + info->dlpi_addr); for (j = 0; j < num_dyn; j++) { if (*fn && *num_fn) break; if (dyn->d_tag == DT_NULL) { break; } else if (dyn->d_tag == tag_a) { *fn = (void (**)(void))(dyn->d_un.d_ptr + info->dlpi_addr); } else if (dyn->d_tag == tag_s) { *num_fn = dyn->d_un.d_val / sizeof(Elf_Addr); } dyn++; } } } void __utee_call_elf_init_fn(void) { void (**fn)(void) = NULL; size_t num_mod = 0; size_t num_fn = 0; size_t mod = 0; size_t i = 0; struct phdr_info *qe = NULL; struct phdr_info *qe2 = NULL; dl_iterate_phdr(_init_iterate_phdr_cb, &num_mod); /* Reverse order: dependencies first */ TAILQ_FOREACH_REVERSE(qe, &__phdr_info_head, phdr_info_head, link) { if (mod == num_mod - _num_mod_init) break; _get_fn_array(&qe->info, DT_INIT_ARRAY, DT_INIT_ARRAYSZ, &fn, &num_fn); for (i = 0; i < num_fn; i++) fn[i](); fn = NULL; num_fn = 0; mod++; } _num_mod_init += mod; TAILQ_FOREACH_SAFE(qe, &__phdr_info_head, link, qe2) { TAILQ_REMOVE(&__phdr_info_head, qe, link); free(qe); } } static int _fini_iterate_phdr_cb(struct dl_phdr_info *info, size_t size __unused, void *data __unused) { void (**fn)(void) = NULL; size_t num_fn = 0; size_t i = 0; _get_fn_array(info, DT_FINI_ARRAY, DT_FINI_ARRAYSZ, &fn, &num_fn); for (i = 1; i <= num_fn; i++) fn[num_fn - i](); return 0; } void __utee_call_elf_fini_fn(void) { dl_iterate_phdr(_fini_iterate_phdr_cb, NULL); } static TEE_Result init_instance(void) { trace_set_level(tahead_get_trace_level()); __utee_gprof_init(); malloc_add_pool(ta_heap, ta_heap_size); _TEE_MathAPI_Init(); __utee_tcb_init(); __utee_call_elf_init_fn(); return TA_CreateEntryPoint(); } static void uninit_instance(void) { __utee_gprof_fini(); TA_DestroyEntryPoint(); __utee_call_elf_fini_fn(); } static void ta_header_save_params(uint32_t param_types, TEE_Param params[TEE_NUM_PARAMS]) { ta_param_types = param_types; if (params) memcpy(ta_params, params, sizeof(ta_params)); else memset(ta_params, 0, sizeof(ta_params)); } static struct ta_session *ta_header_get_session(uint32_t session_id) { struct ta_session *itr; TAILQ_FOREACH(itr, &ta_sessions, link) { if (itr->session_id == session_id) return itr; } return NULL; } static TEE_Result ta_header_add_session(uint32_t session_id) { struct ta_session *itr = ta_header_get_session(session_id); TEE_Result res; if (itr) return TEE_SUCCESS; if (!init_done) { init_done = true; res = init_instance(); if (res) return res; } itr = TEE_Malloc(sizeof(struct ta_session), TEE_USER_MEM_HINT_NO_FILL_ZERO); if (!itr) return TEE_ERROR_OUT_OF_MEMORY; itr->session_id = session_id; itr->session_ctx = 0; TAILQ_INSERT_TAIL(&ta_sessions, itr, link); return TEE_SUCCESS; } static void ta_header_remove_session(uint32_t session_id) { struct ta_session *itr; bool keep_alive; TAILQ_FOREACH(itr, &ta_sessions, link) { if (itr->session_id == session_id) { TAILQ_REMOVE(&ta_sessions, itr, link); TEE_Free(itr); keep_alive = (ta_head.flags & TA_FLAG_SINGLE_INSTANCE) && (ta_head.flags & TA_FLAG_INSTANCE_KEEP_ALIVE); if (TAILQ_EMPTY(&ta_sessions) && !keep_alive) uninit_instance(); return; } } } static void to_utee_params(struct utee_params *up, uint32_t param_types, const TEE_Param params[TEE_NUM_PARAMS]) { size_t n = 0; up->types = param_types; for (n = 0; n < TEE_NUM_PARAMS; n++) { switch (TEE_PARAM_TYPE_GET(param_types, n)) { case TEE_PARAM_TYPE_VALUE_INPUT: case TEE_PARAM_TYPE_VALUE_OUTPUT: case TEE_PARAM_TYPE_VALUE_INOUT: up->vals[n * 2] = params[n].value.a; up->vals[n * 2 + 1] = params[n].value.b; break; case TEE_PARAM_TYPE_MEMREF_INPUT: case TEE_PARAM_TYPE_MEMREF_OUTPUT: case TEE_PARAM_TYPE_MEMREF_INOUT: up->vals[n * 2] = (uintptr_t)params[n].memref.buffer; up->vals[n * 2 + 1] = params[n].memref.size; break; default: up->vals[n * 2] = 0; up->vals[n * 2 + 1] = 0; break; } } } static void from_utee_params(TEE_Param params[TEE_NUM_PARAMS], uint32_t *param_types, const struct utee_params *up) { size_t n; uint32_t types = up->types; for (n = 0; n < TEE_NUM_PARAMS; n++) { uintptr_t a = up->vals[n * 2]; uintptr_t b = up->vals[n * 2 + 1]; switch (TEE_PARAM_TYPE_GET(types, n)) { case TEE_PARAM_TYPE_VALUE_INPUT: case TEE_PARAM_TYPE_VALUE_OUTPUT: case TEE_PARAM_TYPE_VALUE_INOUT: params[n].value.a = a; params[n].value.b = b; break; case TEE_PARAM_TYPE_MEMREF_INPUT: case TEE_PARAM_TYPE_MEMREF_OUTPUT: case TEE_PARAM_TYPE_MEMREF_INOUT: params[n].memref.buffer = (void *)a; params[n].memref.size = b; break; default: break; } } if (param_types) *param_types = types; } static TEE_Result entry_open_session(unsigned long session_id, struct utee_params *up) { TEE_Result res; struct ta_session *session; uint32_t param_types; TEE_Param params[TEE_NUM_PARAMS]; res = ta_header_add_session(session_id); if (res != TEE_SUCCESS) return res; session = ta_header_get_session(session_id); if (!session) return TEE_ERROR_BAD_STATE; from_utee_params(params, &param_types, up); ta_header_save_params(param_types, params); res = TA_OpenSessionEntryPoint(param_types, params, &session->session_ctx); to_utee_params(up, param_types, params); if (res != TEE_SUCCESS) ta_header_remove_session(session_id); return res; } static TEE_Result entry_close_session(unsigned long session_id) { struct ta_session *session = ta_header_get_session(session_id); if (!session) return TEE_ERROR_BAD_STATE; TA_CloseSessionEntryPoint(session->session_ctx); ta_header_remove_session(session_id); return TEE_SUCCESS; } static TEE_Result entry_invoke_command(unsigned long session_id, struct utee_params *up, unsigned long cmd_id) { TEE_Result res; uint32_t param_types; TEE_Param params[TEE_NUM_PARAMS]; struct ta_session *session = ta_header_get_session(session_id); if (!session) return TEE_ERROR_BAD_STATE; from_utee_params(params, &param_types, up); ta_header_save_params(param_types, params); res = TA_InvokeCommandEntryPoint(session->session_ctx, cmd_id, param_types, params); to_utee_params(up, param_types, params); return res; } TEE_Result __utee_entry(unsigned long func, unsigned long session_id, struct utee_params *up, unsigned long cmd_id) { TEE_Result res; switch (func) { case UTEE_ENTRY_FUNC_OPEN_SESSION: res = entry_open_session(session_id, up); break; case UTEE_ENTRY_FUNC_CLOSE_SESSION: res = entry_close_session(session_id); break; case UTEE_ENTRY_FUNC_INVOKE_COMMAND: res = entry_invoke_command(session_id, up, cmd_id); break; default: res = 0xffffffff; TEE_Panic(0); break; } ta_header_save_params(0, NULL); return res; }
4,063
381
<reponame>joyrun/ActivityRouter package com.grouter; import java.lang.annotation.ElementType; import java.lang.annotation.Retention; import java.lang.annotation.RetentionPolicy; import java.lang.annotation.Target; /** * 默认使用空构造方法实现,如果带有参数的构造方法需要实现就要使用ComponentConstructor注解 * Created by Wiki on 19/8/14. */ @Target({ElementType.TYPE}) @Retention(RetentionPolicy.RUNTIME) public @interface RouterComponent { /** * path */ String value() default ""; /** * 接口 */ Class protocol(); /** * 用于文档,名称 */ String name() default ""; /** * 用于文档,描述 */ String description() default ""; }
341
190,993
/* Copyright 2017 The TensorFlow Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ==============================================================================*/ #include <cmath> #include "tensorflow/core/platform/logging.h" #include "tensorflow/lite/toco/graph_transformations/graph_transformations.h" #include "tensorflow/lite/toco/model.h" #include "tensorflow/lite/toco/tooling_util.h" namespace toco { template <ArrayDataType A, typename T> void FillRangeOutput(const Array& start_array, const Array& limit_array, const Array& delta_array, Array* output_array) { // Compute buffer contents T start = start_array.GetBuffer<A>().data[0]; T limit = limit_array.GetBuffer<A>().data[0]; T delta = delta_array.GetBuffer<A>().data[0]; auto& buffer = output_array->GetMutableBuffer<A>(); buffer.data.clear(); int size = (std::is_integral<T>::value ? ((std::abs(limit - start) + std::abs(delta) - 1) / std::abs(delta)) : std::ceil(std::abs((limit - start) / delta))); for (int i = 0; i < size; ++i) { buffer.data.push_back(start + i * delta); } CHECK_EQ(std::floor((limit - start) / delta), buffer.data.size()); CHECK_EQ(buffer.data.size(), output_array->shape().dims()[0]); } ::tensorflow::Status ResolveConstantRange::Run(Model* model, std::size_t op_index, bool* modified) { *modified = false; const auto it = model->operators.begin() + op_index; auto* base_op = it->get(); if (base_op->type != OperatorType::kRange) { return ::tensorflow::Status::OK(); } auto* op = static_cast<RangeOperator*>(base_op); CHECK_EQ(op->inputs.size(), 3); const auto& start_array = model->GetArray(op->inputs[0]); if (!start_array.has_shape()) { // Yield until all input dims have been resolved. return ::tensorflow::Status::OK(); } const auto& limit_array = model->GetArray(op->inputs[1]); if (!limit_array.has_shape()) { // Yield until all input dims have been resolved. return ::tensorflow::Status::OK(); } const auto& delta_array = model->GetArray(op->inputs[2]); if (!delta_array.has_shape()) { // Yield until all input dims have been resolved. return ::tensorflow::Status::OK(); } for (const auto& input : op->inputs) { if (!IsConstantParameterArray(*model, input)) { // yield if any input is mutable return ::tensorflow::Status::OK(); } } CHECK_EQ(op->outputs.size(), 1); auto& output_array = model->GetArray(op->outputs[0]); if (output_array.data_type == ArrayDataType::kNone) { // Yield until the output type has been set by PropagateArrayDataTypes return ::tensorflow::Status::OK(); } CHECK_EQ(RequiredBufferSizeForShape(start_array.shape()), 1) << "Range op inputs must be scalar."; CHECK_EQ(RequiredBufferSizeForShape(limit_array.shape()), 1) << "Range op inputs must be scalar."; CHECK_EQ(RequiredBufferSizeForShape(delta_array.shape()), 1) << "Range op inputs must be scalar."; CHECK(start_array.data_type == ArrayDataType::kInt32 || start_array.data_type == ArrayDataType::kFloat) << "Range op inputs must be int32 or float."; CHECK(limit_array.data_type == start_array.data_type) << "Range op inputs type must be equal."; CHECK(delta_array.data_type == start_array.data_type) << "Range op inputs type must be equal."; if (start_array.data_type == ArrayDataType::kInt32) { FillRangeOutput<ArrayDataType::kInt32, int32_t>(start_array, limit_array, delta_array, &output_array); } else { FillRangeOutput<ArrayDataType::kFloat, float>(start_array, limit_array, delta_array, &output_array); } DeleteOpAndArrays(model, op); *modified = true; return ::tensorflow::Status::OK(); } } // namespace toco
1,692
1,679
<filename>src/boards/mcu/saml21/hal/include/hpl_gpio.h /** * \file * * \brief Port related functionality declaration. * * Copyright (C) 2014-2016 Atmel Corporation. All rights reserved. * * \asf_license_start * * \page License * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, * this list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * 3. The name of Atmel may not be used to endorse or promote products derived * from this software without specific prior written permission. * * 4. This software may only be redistributed and used in connection with an * Atmel microcontroller product. * * THIS SOFTWARE IS PROVIDED BY ATMEL "AS IS" AND ANY EXPRESS OR IMPLIED * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT ARE * EXPRESSLY AND SPECIFICALLY DISCLAIMED. IN NO EVENT SHALL ATMEL BE LIABLE FOR * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. * * \asf_license_stop * */ #ifndef _HPL_GPIO_H_INCLUDED #define _HPL_GPIO_H_INCLUDED /** * \addtogroup HPL Port * * \section hpl_port_rev Revision History * - v1.0.0 Initial Release * *@{ */ #include <compiler.h> #ifdef __cplusplus extern "C" { #endif /** * \brief Macros for the pin and port group, lower 5 * bits stands for pin number in the group, higher 3 * bits stands for port group */ #define GPIO_PIN(n) (((n)&0x1Fu) << 0) #define GPIO_PORT(n) ((n) >> 5) #define GPIO(port, pin) ((((port)&0x7u) << 5) + ((pin)&0x1Fu)) #define GPIO_PIN_FUNCTION_OFF 0xffffffff /** * \brief PORT pull mode settings */ enum gpio_pull_mode { GPIO_PULL_OFF, GPIO_PULL_UP, GPIO_PULL_DOWN }; /** * \brief PORT direction settins */ enum gpio_direction { GPIO_DIRECTION_OFF, GPIO_DIRECTION_IN, GPIO_DIRECTION_OUT }; /** * \brief PORT group abstraction */ enum gpio_port { GPIO_PORTA, GPIO_PORTB, GPIO_PORTC, GPIO_PORTD, GPIO_PORTE }; /** * \name HPL functions */ //@{ /** * \brief Port initialization function * * Port initialization function should setup the port module based * on a static configuration file, this function should normally * not be called directly, but is a part of hal_init() */ void _gpio_init(void); /** * \brief Set direction on port with mask * * Set data direction for each pin, or disable the pin * * \param[in] port Ports are grouped into groups of maximum 32 pins, * GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc * \param[in] mask Bit mask where 1 means apply direction setting to the * corresponding pin * \param[in] direction GPIO_DIRECTION_OFF = set pin direction to input * and disable input buffer to disable the pin * GPIO_DIRECTION_IN = set pin direction to input * and enable input buffer to enable the pin * GPIO_DIRECTION_OUT = set pin direction to output * and disable input buffer */ static inline void _gpio_set_direction(const enum gpio_port port, const uint32_t mask, const enum gpio_direction direction); /** * \brief Set output level on port with mask * * Sets output state on pin to high or low with pin masking * * \param[in] port Ports are grouped into groups of maximum 32 pins, * GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc * \param[in] mask Bit mask where 1 means apply direction setting to * the corresponding pin * \param[in] level true = pin level is set to 1 * false = pin level is set to 0 */ static inline void _gpio_set_level(const enum gpio_port port, const uint32_t mask, const bool level); /** * \brief Change output level to the opposite with mask * * Change pin output level to the opposite with pin masking * * \param[in] port Ports are grouped into groups of maximum 32 pins, * GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc * \param[in] mask Bit mask where 1 means apply direction setting to * the corresponding pin */ static inline void _gpio_toggle_level(const enum gpio_port port, const uint32_t mask); /** * \brief Get input levels on all port pins * * Get input level on all port pins, will read IN register if configured to * input and OUT register if configured as output * * \param[in] port Ports are grouped into groups of maximum 32 pins, * GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc */ static inline uint32_t _gpio_get_level(const enum gpio_port port); /** * \brief Set pin pull mode * * Set pull mode on a single pin * * \notice This function will automatically change pin direction to input * * \param[in] port Ports are grouped into groups of maximum 32 pins, * GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc * \param[in] pin The pin in the group that pull mode should be selected * for * \param[in] pull_mode GPIO_PULL_OFF = pull resistor on pin is disabled * GPIO_PULL_DOWN = pull resistor on pin will pull pin * level to ground level * GPIO_PULL_UP = pull resistor on pin will pull pin * level to VCC */ static inline void _gpio_set_pin_pull_mode(const enum gpio_port port, const uint8_t pin, const enum gpio_pull_mode pull_mode); /** * \brief Set gpio function * * Select which function a gpio is used for * * \param[in] gpio The gpio to set function for * \param[in] function The gpio function is given by a 32-bit wide bitfield * found in the header files for the device * */ static inline void _gpio_set_pin_function(const uint32_t gpio, const uint32_t function); #include <hpl_gpio_base.h> //@} #ifdef __cplusplus } #endif /**@}*/ #endif /* _HPL_GPIO_H_INCLUDED */
2,449
487
print( f"{dumps(data)}\n" )
14
1,355
<gh_stars>1000+ // Copyright (c) 2018 <NAME> // // I am making my contributions/submissions to this project solely in my // personal capacity and am not conveying any rights to any intellectual // property of any third parties. #include "points_to_implicit.h" #include "pybind11_utils.h" #include <jet/points_to_implicit2.h> #include <jet/points_to_implicit3.h> namespace py = pybind11; using namespace jet; void addPointsToImplicit2(pybind11::module& m) { py::class_<PointsToImplicit2, PointsToImplicit2Ptr>(m, "PointsToImplicit2", R"pbdoc( Abstract base class for 2-D points-to-implicit converters. )pbdoc") .def("convert", [](const PointsToImplicit2& instance, const py::list& points, ScalarGrid2Ptr output) { std::vector<Vector2D> points_; for (size_t i = 0; i < points.size(); ++i) { points_.push_back(objectToVector2D(points[i])); } ConstArrayAccessor1<Vector2D> pointsAcc(points_.size(), points_.data()); instance.convert(pointsAcc, output.get()); }, R"pbdoc( Converts the given points to implicit surface scalar field. Parameters ---------- - points : List of 2D vectors. - output : Scalar grid output. )pbdoc", py::arg("points"), py::arg("output")); } void addPointsToImplicit3(pybind11::module& m) { py::class_<PointsToImplicit3, PointsToImplicit3Ptr>(m, "PointsToImplicit3", R"pbdoc( Abstract base class for 3-D points-to-implicit converters. )pbdoc") .def("convert", [](const PointsToImplicit3& instance, const py::list& points, ScalarGrid3Ptr output) { std::vector<Vector3D> points_; for (size_t i = 0; i < points.size(); ++i) { points_.push_back(objectToVector3D(points[i])); } ConstArrayAccessor1<Vector3D> pointsAcc(points_.size(), points_.data()); instance.convert(pointsAcc, output.get()); }, R"pbdoc( Converts the given points to implicit surface scalar field. Parameters ---------- - points : List of 3D vectors. - output : Scalar grid output. )pbdoc", py::arg("points"), py::arg("output")); }
1,400
1,279
/* * Copyright (c) 2012-2021 <NAME> et al. * License: https://github.com/dbartolini/crown/blob/master/LICENSE */ #pragma once #include "core/filesystem/types.h" #include "core/memory/types.h" #include "core/types.h" #include "resource/types.h" namespace crown { namespace config_resource_internal { s32 compile(CompileOptions& opts); void* load(File& file, Allocator& a); void unload(Allocator& allocator, void* resource); } // namespace config_resource_internal } // namespace crown
177
558
# Copyright 2018 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may not # use this file except in compliance with the License. You may obtain a copy of # the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations under # the License. # ============================================================================== """test ffmpeg dataset""" import os import sys import pytest import tensorflow as tf import tensorflow_io as tfio @pytest.fixture(name="video_data", scope="module") def fixture_video_data(): """fixture_video_data""" path = "file://" + os.path.join( os.path.dirname(os.path.abspath(__file__)), "test_video", "small.mp4" ) # TODO: get raw value value = tf.zeros((166, 320, 560, 3), tf.uint8) return path, value @pytest.mark.parametrize( ("io_dataset_func"), [ pytest.param(lambda f: tfio.IODataset.graph(tf.uint8).from_ffmpeg(f, "v:0")), pytest.param(lambda f: tfio.IODataset.from_ffmpeg(f, "v:0")), ], ids=["from_ffmpeg", "from_ffmpeg(eager)"], ) def test_video_io_dataset(video_data, io_dataset_func): """test_video_io_dataset""" video_path, video_value = video_data video_dataset = io_dataset_func(video_path) i = 0 for value in video_dataset: assert video_value[i].shape == value.shape i += 1 assert i == 166 video_dataset = io_dataset_func(video_path).batch(2) i = 0 for value in video_dataset: assert video_value[i : i + 2].shape == value.shape i += 2 assert i == 166 @pytest.mark.parametrize( ("io_dataset_func"), [pytest.param(lambda f: tfio.IODataset.graph(tf.uint8).from_ffmpeg(f, "v:0"),)], ids=["from_ffmpeg"], ) def test_video_io_dataset_with_dataset(video_data, io_dataset_func): """test_video_io_dataset_with_dataset""" video_path, video_value = video_data filename_dataset = tf.data.Dataset.from_tensor_slices([video_path, video_path]) position_dataset = tf.data.Dataset.from_tensor_slices( [tf.constant(50, tf.int64), tf.constant(100, tf.int64)] ) dataset = tf.data.Dataset.zip((filename_dataset, position_dataset)) # Note: @tf.function is actually not needed, as tf.data.Dataset # will automatically wrap the `func` into a graph anyway. # The following is purely for explanation purposes. # Return: an embedded dataset (in an outer dataset) for position:position+100 @tf.function def func(filename, position): video_dataset = io_dataset_func(filename) return video_dataset.skip(position).take(10) dataset = dataset.map(func) item = 0 # Notice video_dataset in dataset: for video_dataset in dataset: position = 50 if item == 0 else 100 i = 0 for value in video_dataset: assert video_value[position + i].shape == value.shape i += 1 assert i == 10 item += 1 assert item == 2
1,292
310
<filename>gear/hardware/0/600p-crl.json { "name": "600P CRL", "description": "A glass cutter.", "url": "https://www.amazon.com/LAURENCE-Heavy-Duty-Oil-Type-Pattern-Cutting/dp/B006JFN2OG" }
83
309
// Copyright 2019 the deepx authors. // Author: <NAME> (<EMAIL>) // #include <deepx_core/graph/op_impl.h> namespace deepx_core { namespace { bool SequenceMaskInferShape(const Shape& X, int max_size, Shape* Z) noexcept { if (!X.is_rank(1)) { DXERROR("Invalid X: rank of X %d must be 1.", X.rank()); return false; } if (max_size <= 0) { DXERROR("Invalid max_size: max_size %d must be positive.", max_size); return false; } Z->resize(X[0], max_size); return true; } template <typename T> void SequenceMask(const Tensor<T>& X, Tensor<T>* Z) noexcept { const T* _X = X.data(); T* _Z = Z->data(); int batch = Z->dim(0); int m = Z->dim(1); int Xi; Z->zeros(); for (int i = 0; i < batch; ++i) { Xi = (int)*_X; if (Xi > 0) { if (Xi > m) { Xi = m; } for (int j = 0; j < Xi; ++j) { _Z[j] = 1; } } _X += 1; _Z += m; } } } // namespace SequenceMaskNode::SequenceMaskNode(std::string name, GraphNode* X, int max_size) : GraphNodeUnaryBase(std::move(name), X), max_size_(max_size) { if (!X->shape().empty()) { (void)SequenceMaskInferShape(X->shape(), max_size_, &shape_); } } class SequenceMaskOp : public OpUnaryBase { private: Shape Zshape_; public: DEFINE_OP_LIKE(SequenceMaskOp); const Shape& InferShape() override { int max_size = ((const SequenceMaskNode*)node_)->max_size(); DXCHECK_THROW(SequenceMaskInferShape(X_->shape(), max_size, &Zshape_)); return Zshape_; } void Forward() override { SequenceMask(*X_, Z_); } }; GRAPH_NODE_OP_REGISTER(SequenceMask); } // namespace deepx_core
697
665
<filename>3D_Printed_LED_Microphone_Flag/code.py<gh_stars>100-1000 # SPDX-FileCopyrightText: 2013 <NAME> for Adafruit Industries # SPDX-FileCopyrightText: 2017 Mi<NAME> for Adafruit Industries # # SPDX-License-Identifier: BSD # LED VU meter for Arduino and Adafruit NeoPixel LEDs. # Hardware requirements: # - M0 boards # - Adafruit Electret Microphone Amplifier (ID: 1063) # - Adafruit Flora RGB Smart Pixels (ID: 1260) # OR # - Adafruit NeoPixel Digital LED strip (ID: 1138) # - Optional: battery for portable use (else power through USB or adapter) # Software requirements: # - Adafruit NeoPixel library # Connections: # - 3.3V to mic amp + # - GND to mic amp - # - Analog pin to microphone output (configurable below) # - Digital pin to LED data input (configurable below) # See notes in setup() regarding 5V vs. 3.3V boards - there may be an # extra connection to make and one line of code to enable or disable. # Written by Adafruit Industries. Distributed under the BSD license. # This paragraph must be included in any redistribution. # fscale function: # Floating Point Autoscale Function V0.1 # Written by <NAME> 2007 # Modified fromhere code by <NAME> # Ported to Circuit Python by <NAME> import time import board import neopixel from rainbowio import colorwheel from analogio import AnalogIn n_pixels = 16 # Number of pixels you are using mic_pin = AnalogIn(board.A1) # Microphone is attached to this analog pin led_pin = board.D1 # NeoPixel LED strand is connected to this pin sample_window = .1 # Sample window for average level peak_hang = 24 # Time of pause before peak dot falls peak_fall = 4 # Rate of falling peak dot input_floor = 10 # Lower range of analogRead input # Max range of analogRead input, the lower the value the more sensitive # (1023 = max) input_ceiling = 300 peak = 16 # Peak level of column; used for falling dots sample = 0 dotcount = 0 # Frame counter for peak dot dothangcount = 0 # Frame counter for holding peak dot strip = neopixel.NeoPixel(led_pin, n_pixels, brightness=1, auto_write=False) def remapRange(value, leftMin, leftMax, rightMin, rightMax): # this remaps a value fromhere original (left) range to new (right) range # Figure out how 'wide' each range is leftSpan = leftMax - leftMin rightSpan = rightMax - rightMin # Convert the left range into a 0-1 range (int) valueScaled = int(value - leftMin) / int(leftSpan) # Convert the 0-1 range into a value in the right range. return int(rightMin + (valueScaled * rightSpan)) def fscale(originalmin, originalmax, newbegin, newend, inputvalue, curve): invflag = 0 # condition curve parameter # limit range if curve > 10: curve = 10 if curve < -10: curve = -10 # - invert and scale - # this seems more intuitive # postive numbers give more weight to high end on output curve = (curve * -.1) # convert linear scale into lograthimic exponent for other pow function curve = pow(10, curve) # Check for out of range inputValues if inputvalue < originalmin: inputvalue = originalmin if inputvalue > originalmax: inputvalue = originalmax # Zero Refference the values originalrange = originalmax - originalmin if newend > newbegin: newrange = newend - newbegin else: newrange = newbegin - newend invflag = 1 zerorefcurval = inputvalue - originalmin # normalize to 0 - 1 float normalizedcurval = zerorefcurval / originalrange # Check for originalMin > originalMax # -the math for all other cases # i.e. negative numbers seems to work out fine if originalmin > originalmax: return 0 if invflag == 0: rangedvalue = (pow(normalizedcurval, curve) * newrange) + newbegin else: # invert the ranges rangedvalue = newbegin - (pow(normalizedcurval, curve) * newrange) return rangedvalue def drawLine(fromhere, to): if fromhere > to: to, fromhere = fromhere, to for index in range(fromhere, to): strip[index] = (0, 0, 0) while True: time_start = time.monotonic() # current time used for sample window peaktopeak = 0 # peak-to-peak level signalmax = 0 signalmin = 1023 c = 0 y = 0 # collect data for length of sample window (in seconds) while (time.monotonic() - time_start) < sample_window: # convert to arduino 10-bit [1024] fromhere 16-bit [65536] sample = mic_pin.value / 64 if sample < 1024: # toss out spurious readings if sample > signalmax: signalmax = sample # save just the max levels elif sample < signalmin: signalmin = sample # save just the min levels peaktopeak = signalmax - signalmin # max - min = peak-peak amplitude # Fill the strip with rainbow gradient for i in range(0, len(strip)): strip[i] = colorwheel(remapRange(i, 0, (n_pixels - 1), 30, 150)) # Scale the input logarithmically instead of linearly c = fscale(input_floor, input_ceiling, (n_pixels - 1), 0, peaktopeak, 2) if c < peak: peak = c # keep dot on top dothangcount = 0 # make the dot hang before falling if c <= n_pixels: # fill partial column with off pixels drawLine(n_pixels, n_pixels - int(c)) # Set the peak dot to match the rainbow gradient y = n_pixels - peak strip.fill = (y - 1, colorwheel(remapRange(y, 0, (n_pixels - 1), 30, 150))) strip.write() # Frame based peak dot animation if dothangcount > peak_hang: # Peak pause length dotcount += 1 if dotcount >= peak_fall: # Fall rate peak += 1 dotcount = 0 else: dothangcount += 1
2,039
348
<filename>docs/data/t2/011/11335.json<gh_stars>100-1000 {"nom":"Sainte-Colombe-sur-Guette","dpt":"Aude","inscrits":50,"abs":6,"votants":44,"blancs":10,"nuls":1,"exp":33,"res":[{"panneau":"1","voix":25},{"panneau":"2","voix":8}]}
101
891
from typing import ClassVar from datamodel_code_generator.model.pydantic.base_model import BaseModel class CustomRootType(BaseModel): TEMPLATE_FILE_PATH: ClassVar[str] = 'pydantic/BaseModel_root.jinja2' BASE_CLASS: ClassVar[str] = 'pydantic.BaseModel'
96
2,293
<gh_stars>1000+ /***************************************************************************** * * * This file is part of the BeanShell Java Scripting distribution. * * Documentation and updates may be found at http://www.beanshell.org/ * * * * Sun Public License Notice: * * * * The contents of this file are subject to the Sun Public License Version * * 1.0 (the "License"); you may not use this file except in compliance with * * the License. A copy of the License is available at http://www.sun.com * * * * The Original Code is BeanShell. The Initial Developer of the Original * * Code is <NAME>. Portions created by <NAME> are Copyright * * (C) 2000. All Rights Reserved. * * * * GNU Public License Notice: * * * * Alternatively, the contents of this file may be used under the terms of * * the GNU Lesser General Public License (the "LGPL"), in which case the * * provisions of LGPL are applicable instead of those above. If you wish to * * allow use of your version of this file only under the terms of the LGPL * * and not to allow others to use your version of this file under the SPL, * * indicate your decision by deleting the provisions above and replace * * them with the notice and other provisions required by the LGPL. If you * * do not delete the provisions above, a recipient may use your version of * * this file under either the SPL or the LGPL. * * * * <NAME> (<EMAIL>) * * Author of Learning Java, O'Reilly & Associates * * http://www.pat.net/~pat/ * * * *****************************************************************************/ package bsh; import java.io.*; import java.net.*; import java.text.*; /** Remote executor class. Posts a script from the command line to a BshServlet or embedded interpreter using (respectively) HTTP or the bsh telnet service. Output is printed to stdout and a numeric return value is scraped from the result. */ public class Remote { public static void main( String args[] ) throws Exception { if ( args.length < 2 ) { System.out.println( "usage: Remote URL(http|bsh) file [ file ] ... "); System.exit(1); } String url = args[0]; String text = getFile(args[1]); int ret = eval( url, text ); System.exit( ret ); } /** Evaluate text in the interpreter at url, returning a possible integer return value. */ public static int eval( String url, String text ) throws IOException { String returnValue = null; if ( url.startsWith( "http:" ) ) { returnValue = doHttp( url, text ); } else if ( url.startsWith( "bsh:" ) ) { returnValue = doBsh( url, text ); } else throw new IOException( "Unrecognized URL type." +"Scheme must be http:// or bsh://"); try { return Integer.parseInt( returnValue ); } catch ( Exception e ) { // this convention may change... return 0; } } static String doBsh( String url, String text ) { OutputStream out; InputStream in; String host = ""; String port = ""; String returnValue = "-1"; String orgURL = url; // Need some format checking here try { url = url.substring(6); // remove the bsh:// // get the index of the : between the host and the port is located int index = url.indexOf(":"); host = url.substring(0,index); port = url.substring(index+1,url.length()); } catch ( Exception ex ) { System.err.println("Bad URL: "+orgURL+": "+ex ); return returnValue; } try { System.out.println("Connecting to host : " + host + " at port : " + port); Socket s = new Socket(host, Integer.parseInt(port) + 1); out = s.getOutputStream(); in = s.getInputStream(); sendLine( text, out ); BufferedReader bin = new BufferedReader( new InputStreamReader(in)); String line; while ( (line=bin.readLine()) != null ) System.out.println( line ); // Need to scrape a value from the last line? returnValue="1"; return returnValue; } catch(Exception ex) { System.err.println("Error communicating with server: "+ex); return returnValue; } } private static void sendLine( String line, OutputStream outPipe ) throws IOException { outPipe.write( line.getBytes() ); outPipe.flush(); } /* TODO: this is not unicode friendly, nor is getFile() The output is urlencoded 8859_1 text. should probably be urlencoded UTF-8... how does the servlet determine the encoded charset? I guess we're supposed to add a ";charset" clause to the content type? */ static String doHttp( String postURL, String text ) { String returnValue = null; StringBuffer sb = new StringBuffer(); sb.append( "bsh.client=Remote" ); sb.append( "&bsh.script=" ); sb.append( URLEncoder.encode( text ) ); /* // This requires Java 1.3 try { sb.append( URLEncoder.encode( text, "8859_1" ) ); } catch ( UnsupportedEncodingException e ) { e.printStackTrace(); } */ String formData = sb.toString( ); try { URL url = new URL( postURL ); HttpURLConnection urlcon = (HttpURLConnection) url.openConnection( ); urlcon.setRequestMethod("POST"); urlcon.setRequestProperty("Content-type", "application/x-www-form-urlencoded"); urlcon.setDoOutput(true); urlcon.setDoInput(true); PrintWriter pout = new PrintWriter( new OutputStreamWriter( urlcon.getOutputStream(), "8859_1"), true ); pout.print( formData ); pout.flush(); // read results... int rc = urlcon.getResponseCode(); if ( rc != HttpURLConnection.HTTP_OK ) System.out.println("Error, HTTP response: "+rc ); returnValue = urlcon.getHeaderField("Bsh-Return"); BufferedReader bin = new BufferedReader( new InputStreamReader( urlcon.getInputStream() ) ); String line; while ( (line=bin.readLine()) != null ) System.out.println( line ); System.out.println( "Return Value: "+returnValue ); } catch (MalformedURLException e) { System.out.println(e); // bad postURL } catch (IOException e2) { System.out.println(e2); // I/O error } return returnValue; } /* Note: assumes default character encoding */ static String getFile( String name ) throws FileNotFoundException, IOException { StringBuffer sb = new StringBuffer(); BufferedReader bin = new BufferedReader( new FileReader( name ) ); String line; while ( (line=bin.readLine()) != null ) sb.append( line ).append( "\n" ); return sb.toString(); } }
3,130
653
//==---------------- opencl.hpp - SYCL OpenCL backend ----------------------==// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// #pragma once #include <CL/sycl/accessor.hpp> #include <CL/sycl/backend.hpp> #include <CL/sycl/backend_types.hpp> #include <CL/sycl/detail/backend_traits.hpp> #include <CL/sycl/detail/cl.h> #include <CL/sycl/kernel_bundle.hpp> #include <vector> __SYCL_INLINE_NAMESPACE(cl) { namespace sycl { template <> struct interop<backend::opencl, platform> { using type = cl_platform_id; }; template <> struct interop<backend::opencl, device> { using type = cl_device_id; }; template <> struct interop<backend::opencl, context> { using type = cl_context; }; template <> struct interop<backend::opencl, queue> { using type = cl_command_queue; }; #ifdef __SYCL_INTERNAL_API template <> struct interop<backend::opencl, program> { using type = cl_program; }; #endif template <> struct interop<backend::opencl, event> { using type = cl_event; }; template <typename DataT, int Dimensions, access::mode AccessMode> struct interop<backend::opencl, accessor<DataT, Dimensions, AccessMode, access::target::device, access::placeholder::false_t>> { using type = cl_mem; }; template <typename DataT, int Dimensions, access::mode AccessMode> struct interop<backend::opencl, accessor<DataT, Dimensions, AccessMode, access::target::constant_buffer, access::placeholder::false_t>> { using type = cl_mem; }; template <typename DataT, int Dimensions, access::mode AccessMode> struct interop<backend::opencl, accessor<DataT, Dimensions, AccessMode, access::target::image, access::placeholder::false_t>> { using type = cl_mem; }; template <typename DataT, int Dimensions, typename AllocatorT> struct interop<backend::opencl, buffer<DataT, Dimensions, AllocatorT>> { using type = cl_mem; }; namespace detail { template <bundle_state State> struct BackendInput<backend::opencl, kernel_bundle<State>> { using type = cl_program; }; template <bundle_state State> struct BackendReturn<backend::opencl, kernel_bundle<State>> { using type = std::vector<cl_program>; }; template <> struct BackendInput<backend::opencl, kernel> { using type = cl_kernel; }; template <> struct BackendReturn<backend::opencl, kernel> { using type = cl_kernel; }; template <> struct InteropFeatureSupportMap<backend::opencl> { static constexpr bool MakePlatform = true; static constexpr bool MakeDevice = true; static constexpr bool MakeContext = true; static constexpr bool MakeQueue = true; static constexpr bool MakeEvent = true; static constexpr bool MakeBuffer = true; static constexpr bool MakeKernel = true; static constexpr bool MakeKernelBundle = true; }; } // namespace detail namespace opencl { // Implementation of various "make" functions resides in SYCL RT because // creating SYCL objects requires knowing details not acessible here. // Note that they take opaque pi_native_handle that real OpenCL handles // are casted to. // __SYCL_EXPORT platform make_platform(pi_native_handle NativeHandle); __SYCL_EXPORT device make_device(pi_native_handle NativeHandle); __SYCL_EXPORT context make_context(pi_native_handle NativeHandle); #ifdef __SYCL_INTERNAL_API __SYCL_EXPORT program make_program(const context &Context, pi_native_handle NativeHandle); #endif __SYCL_EXPORT queue make_queue(const context &Context, pi_native_handle InteropHandle); // Construction of SYCL platform. template <typename T, typename detail::enable_if_t< std::is_same<T, platform>::value> * = nullptr> T make(typename interop<backend::opencl, T>::type Interop) { return make_platform(detail::pi::cast<pi_native_handle>(Interop)); } // Construction of SYCL device. template <typename T, typename detail::enable_if_t< std::is_same<T, device>::value> * = nullptr> T make(typename interop<backend::opencl, T>::type Interop) { return make_device(detail::pi::cast<pi_native_handle>(Interop)); } // Construction of SYCL context. template <typename T, typename detail::enable_if_t< std::is_same<T, context>::value> * = nullptr> T make(typename interop<backend::opencl, T>::type Interop) { return make_context(detail::pi::cast<pi_native_handle>(Interop)); } // Construction of SYCL program. #ifdef __SYCL_INTERNAL_API template <typename T, typename detail::enable_if_t< std::is_same<T, program>::value> * = nullptr> T make(const context &Context, typename interop<backend::opencl, T>::type Interop) { return make_program(Context, detail::pi::cast<pi_native_handle>(Interop)); } #endif // Construction of SYCL queue. template <typename T, typename detail::enable_if_t< std::is_same<T, queue>::value> * = nullptr> T make(const context &Context, typename interop<backend::opencl, T>::type Interop) { return make_queue(Context, detail::pi::cast<pi_native_handle>(Interop)); } } // namespace opencl } // namespace sycl } // __SYCL_INLINE_NAMESPACE(cl)
2,040