max_stars_count
int64 301
224k
| text
stringlengths 6
1.05M
| token_count
int64 3
727k
|
---|---|---|
634 | <filename>modules/desktop-swt/desktop-swt-ui-impl/src/main/java/org/eclipse/nebula/cwt/svg/SvgGraphic.java
/****************************************************************************
* Copyright (c) 2008, 2009 <NAME>
*
* This program and the accompanying materials
* are made available under the terms of the Eclipse Public License 2.0
* which accompanies this distribution, and is available at
* https://www.eclipse.org/legal/epl-2.0/
*
* SPDX-License-Identifier: EPL-2.0
*
* Contributors:
* <NAME> <<EMAIL>> - initial API and implementation
*****************************************************************************/
package org.eclipse.nebula.cwt.svg;
import org.eclipse.nebula.cwt.svg.SvgPaint.PaintType;
import org.eclipse.swt.SWT;
import org.eclipse.swt.graphics.GC;
import org.eclipse.swt.graphics.Transform;
/**
* A base abstract class for all types of svg elements which can be
* applied (painted) to a graphics context. These may be shapes which
* can be painted directly, or containers which will paint their children.
*/
public abstract class SvgGraphic extends SvgElement {
String title;
String description;
SvgFill fill;
SvgStroke stroke;
SvgTransform transform;
SvgGraphic(SvgContainer container, String id) {
super(container, id);
fill = new SvgFill(this);
stroke = new SvgStroke(this);
}
/**
* Apply this svg graphic to the given graphics context.
* <p>Note that to support the rather abstract structure of svg,
* each time this method is called all transformations and css properties
* to be calculated and applied. If this is a shape, it will be
* painted to the graphics context. Containers will recursively
* make this call on their children.</p>
* @param gc the gc to use in all graphics operations
*/
public abstract void apply(GC gc);
/**
* Returns the value of the <code>desc</code> element that is a child of this svg element.
* If there is no <code>desc</code> element that is a direct decendent of this element, null
* is returned.
* @return the <code>desc</code> of this svg element
*/
public String getDescription() {
return (description == null) ? null : description;
}
SvgFill getFill() {
SvgFill df = new SvgFill(this);
for(SvgElement el : getAncestry(this)) {
if(el instanceof SvgGraphic) {
SvgFill tmp = ((SvgGraphic) el).fill;
if(tmp.type != null) {
df.type = tmp.type;
}
if(tmp.color != null) {
df.color = tmp.color;
}
if(tmp.linkId != null) {
df.linkId = tmp.linkId;
}
if(tmp.opacity != null) {
df.opacity = tmp.opacity;
}
if(tmp.rule != null) {
df.rule = tmp.rule;
}
}
}
if(df.type == null) {
df.type = PaintType.Color;
}
if(df.color == null) {
df.color = 0;
}
if(df.opacity == null) {
df.opacity = 1f;
}
if(df.rule == null) {
df.rule = SWT.FILL_EVEN_ODD;
}
return df;
}
SvgStroke getStroke() {
SvgStroke ds = new SvgStroke(this);
for(SvgElement el : getAncestry(this)) {
if(el instanceof SvgGraphic) {
SvgStroke tmp = ((SvgGraphic) el).stroke;
if(tmp.type != null) {
ds.type = tmp.type;
}
if(tmp.color != null) {
ds.color = tmp.color;
}
if(tmp.linkId != null) {
ds.linkId = tmp.linkId;
}
if(tmp.opacity != null) {
ds.opacity = tmp.opacity;
}
if(tmp.width != null) {
ds.width = tmp.width;
}
if(tmp.lineCap != null) {
ds.lineCap = tmp.lineCap;
}
if(tmp.lineJoin != null) {
ds.lineJoin = tmp.lineJoin;
}
}
}
if(ds.type == null) {
ds.type = PaintType.None;
}
if(ds.type != PaintType.None) {
if(ds.color == null) {
ds.color = 0;
}
if(ds.opacity == null) {
ds.opacity = 1f;
}
if(ds.width == null) {
ds.width = 1f;
}
if(ds.lineCap == null) {
ds.lineCap = SWT.CAP_FLAT;
}
if(ds.lineJoin == null) {
ds.lineJoin = SWT.JOIN_MITER;
}
}
return ds;
}
Transform getTransform(GC gc) {
Transform t = new Transform(gc.getDevice());
gc.getTransform(t);
for(SvgElement el : getAncestry(this)) {
if(el instanceof SvgFragment) {
SvgTransform st = ((SvgFragment) el).boundsTransform;
if(!st.isIdentity()) {
Transform tmp = new Transform(gc.getDevice());
tmp.setElements(st.data[0], st.data[1], st.data[2], st.data[3], st.data[4], st.data[5]);
t.multiply(tmp);
tmp.dispose();
}
} else if(el instanceof SvgGraphic) {
SvgTransform st = ((SvgGraphic) el).transform;
while(st != null) {
if(!st.isIdentity()) {
Transform tmp = new Transform(gc.getDevice());
tmp.setElements(st.data[0], st.data[1], st.data[2], st.data[3], st.data[4], st.data[5]);
t.multiply(tmp);
tmp.dispose();
}
st = st.next;
}
}
}
return t;
}
/**
* Returns the value of the <code>title</code> element that is a child of this svg element.
* If there is no <code>title</code> element that is a direct decendent of this element, null
* is returned.
* @return the <code>title</code> of this svg element
*/
public String getTitle() {
return (title == null) ? null : title;
}
}
| 2,131 |
587 | <gh_stars>100-1000
/*
* Copyright (C) 2013 salesforce.com, inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.auraframework.impl.source.file;
import java.io.File;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
import java.io.IOException;
import java.io.InputStreamReader;
import java.io.Reader;
import java.io.StringReader;
import java.io.StringWriter;
import java.util.Map;
import org.auraframework.def.DefDescriptor;
import org.auraframework.def.Definition;
import org.auraframework.impl.source.AbstractSourceImpl;
import org.auraframework.system.Parser.Format;
import org.auraframework.system.Source;
import org.auraframework.system.TextSource;
import org.auraframework.throwable.AuraRuntimeException;
import org.auraframework.util.IOUtil;
import org.auraframework.util.text.Hash;
import org.auraframework.util.text.HashingReader;
import com.google.common.collect.ImmutableMap;
public class FileSource<D extends Definition> extends AbstractSourceImpl<D> implements TextSource<D> {
private final File file;
private final long lastModified;
private final Hash hash;
private volatile String contents;
/**
* The 'normal' constructor for a file source.
*
* This constructor gets both the file path and the mime type from the file given.
* The file must exist prior to constructing the source. Behaviour is undefined
* if the file does not exist, please do not test this case.
*
* @param descriptor the descriptor for the source.
* @param file the file that contains the source.
*/
public FileSource(DefDescriptor<D> descriptor, File file) throws IOException {
super(descriptor, getFilePath(file), getMimeTypeFromExtension(file.getName()));
this.file = file;
this.lastModified = file.lastModified();
this.hash = Hash.createPromise();
}
/**
* A 'format' based constructor.
*
* This constructor gets both the file path and the mime type from the file given.
* The file must exist prior to constructing the source. Behaviour is undefined
* if the file does not exist, please do not test this case.
*
* @param descriptor the descriptor for the source.
* @param file the file that contains the source.
*/
public FileSource(DefDescriptor<D> descriptor, File file, Format format) {
super(descriptor, getFilePath(file), format);
this.file = file;
this.lastModified = file.lastModified();
this.hash = Hash.createPromise();
}
@Override
public Reader getReader() {
return new StringReader(getContents());
}
@Override
public long getLastModified() {
return lastModified;
}
@Override
public String getHash() {
if (!hash.isSet()) {
throw new RuntimeException("No hash set");
}
return hash.toString();
}
public static String getFilePath(File file) {
try {
if (!file.exists()) {
throw new AuraRuntimeException("File does not exist: " + file.getPath());
}
return file.getCanonicalPath();
} catch (Exception e) {
throw new AuraRuntimeException(e);
}
}
/**
* @see Source#getContents()
*/
@Override
public String getContents() {
if (contents != null) {
return contents;
}
synchronized (this) {
if (contents != null) {
return contents;
}
try {
Reader reader = new InputStreamReader(new FileInputStream(file), "UTF8");
reader = new HashingReader(reader, hash);
StringWriter sw = new StringWriter();
IOUtil.copyStream(reader, sw);
contents = sw.toString();
} catch (FileNotFoundException e) {
throw new AuraRuntimeException(e);
} catch (IOException e) {
throw new AuraRuntimeException(e);
}
}
return contents;
}
private final static Map<String,String> mimeMap;
static {
mimeMap = new ImmutableMap.Builder<String,String>()
.put("js", MIME_JAVASCRIPT)
.put("html", MIME_HTML)
.put("cmp", MIME_XML)
.put("lib", MIME_XML)
.put("app", MIME_XML)
.put("css", MIME_CSS)
.put("evt", MIME_XML)
.put("intf", MIME_XML)
.put("svg", MIME_SVG)
.build();
}
/**
* Get an aura mime type based on the extension.
*
* This is used to figure out mime types for files.
*
* @param name the file name to use.
* @return a mime type for the file, "X-application/unknown" for anything we don't understand.
*/
public static String getMimeTypeFromExtension(String name) {
int idx = name.lastIndexOf('.');
String mimetype = null;
if (idx != -1) {
String ext = name.substring(idx+1);
mimetype = mimeMap.get(ext);
}
if (mimetype != null) {
return mimetype;
}
return "X-application/unknown";
}
}
| 2,299 |
351 | #!/usr/bin/env python3
from tryalgo.dijkstra import dijkstra
from tryalgo.graph import listlist_and_matrix_to_listdict
graph = [[1, 3],
[0, 2, 3],
[1, 4, 5],
[0, 1, 5, 6],
[2, 7],
[2, 3, 7, 8],
[3, 8, 9],
[4, 5, 10],
[5, 6, 10],
[6, 11],
[7, 8],
[9]]
_ = None
# 0 1 2 3 4 5 6 7 8 9 10 11
weights = [[_, 1, _, 4, _, _, _, _, _, _, _, _], # 0
[1, _, 1, 3, _, _, _, _, _, _, _, _], # 1
[_, 1, _, _, 3, 8, _, _, _, _, _, _], # 2
[4, 3, _, _, _, 2, 2, _, _, _, _, _], # 3
[_, _, 3, _, _, _, _, 1, _, _, _, _], # 4
[_, _, 8, 2, _, _, _, 2, 7, _, _, _], # 5
[_, _, _, 2, _, _, _, _, 3, 2, _, _], # 6
[_, _, _, _, 1, 2, _, _, _, _, 3, _], # 7
[_, _, _, _, _, 7, 3, _, _, _, 2, _], # 8
[_, _, _, _, _, _, 2, _, _, _, _, 1], # 9
[_, _, _, _, _, _, _, 3, 2, _, _, _], #10
[_, _, _, _, _, _, _, _, _, 1, _, _]] #11
dist, prec = dijkstra(graph, weights, source=0)
print(dist[10])
print("%i %i %i %i %i %i" % (10, prec[10], prec[prec[10]], prec[prec[prec[10]]],
prec[prec[prec[prec[10]]]], prec[prec[prec[prec[prec[10]]]]]))
sparse_graph = listlist_and_matrix_to_listdict(weights)
# provides the same behavior
dist, prec = dijkstra(graph, weights, source=0)
print(dist[10])
print("%i %i %i %i %i %i" % (10, prec[10], prec[prec[10]], prec[prec[prec[10]]],
prec[prec[prec[prec[10]]]], prec[prec[prec[prec[prec[10]]]]]))
| 959 |
852 | #ifndef DataFormats_ParticleFlowReco_PFCluster_h
#define DataFormats_ParticleFlowReco_PFCluster_h
#include <iostream>
#include <vector>
#include <algorithm>
#include <Rtypes.h>
#include "DataFormats/CaloRecHit/interface/CaloCluster.h"
#include "DataFormats/CaloRecHit/interface/CaloClusterFwd.h"
#include "DataFormats/Math/interface/Point3D.h"
#include "DataFormats/ParticleFlowReco/interface/PFLayer.h"
#include "DataFormats/ParticleFlowReco/interface/PFRecHit.h"
#include "DataFormats/ParticleFlowReco/interface/PFRecHitFraction.h"
#include "Math/GenVector/PositionVector3D.h"
class PFClusterAlgo;
namespace reco {
/**\class PFCluster
\brief Particle flow cluster, see clustering algorithm in PFClusterAlgo
A particle flow cluster is defined by its energy and position, which are
calculated from a vector of PFRecHitFraction. This calculation is
performed in PFClusterAlgo.
\todo Clean up this class to a common base (talk to Paolo Meridiani)
the extra internal stuff (like the vector of PFRecHitFraction's)
could be moved to a PFClusterExtra.
\todo Now that PFRecHitFraction's hold a reference to the PFRecHit's,
put back the calculation of energy and position to PFCluster.
\todo Add an operator+=
\author <NAME>
\date July 2006
*/
class PFCluster : public CaloCluster {
public:
typedef std::vector<std::pair<CaloClusterPtr::key_type, edm::Ptr<PFCluster>>> EEtoPSAssociation;
// Next typedef uses double in ROOT 6 rather than Double32_t due to a bug in ROOT 5,
// which otherwise would make ROOT5 files unreadable in ROOT6. This does not increase
// the size on disk, because due to the bug, double was actually stored on disk in ROOT 5.
typedef ROOT::Math::PositionVector3D<ROOT::Math::CylindricalEta3D<double>> REPPoint;
PFCluster() : CaloCluster(CaloCluster::particleFlow), time_(-99.0), layer_(PFLayer::NONE) {}
/// constructor
PFCluster(PFLayer::Layer layer, double energy, double x, double y, double z);
/// resets clusters parameters
void reset();
/// reset only hits and fractions
void resetHitsAndFractions();
/// add a given fraction of the rechit
void addRecHitFraction(const reco::PFRecHitFraction& frac);
/// vector of rechit fractions
const std::vector<reco::PFRecHitFraction>& recHitFractions() const { return rechits_; }
/// set layer
void setLayer(PFLayer::Layer layer);
/// cluster layer, see PFLayer.h in this directory
PFLayer::Layer layer() const;
/// cluster energy
double energy() const { return energy_; }
/// \return cluster time
float time() const { return time_; }
/// \return the timing uncertainty
float timeError() const { return timeError_; }
/// cluster depth
double depth() const { return depth_; }
void setTime(float time, float timeError = 0) {
time_ = time;
timeError_ = timeError;
}
void setTimeError(float timeError) { timeError_ = timeError; }
void setDepth(double depth) { depth_ = depth; }
/// cluster position: rho, eta, phi
const REPPoint& positionREP() const { return posrep_; }
/// computes posrep_ once and for all
void calculatePositionREP() { posrep_.SetCoordinates(position_.Rho(), position_.Eta(), position_.Phi()); }
/// \todo move to PFClusterTools
static double getDepthCorrection(double energy, bool isBelowPS = false, bool isHadron = false);
PFCluster& operator=(const PFCluster&);
/// some classes to make this fit into a template footprint
/// for RecoPFClusterRefCandidate so we can make jets and MET
/// out of PFClusters.
/// dummy charge
double charge() const { return 0; }
/// transverse momentum, massless approximation
double pt() const { return (energy() * sin(position_.theta())); }
/// angle
double theta() const { return position_.theta(); }
/// dummy vertex access
math::XYZPoint const& vertex() const { return dummyVtx_; }
double vx() const { return vertex().x(); }
double vy() const { return vertex().y(); }
double vz() const { return vertex().z(); }
#if !defined(__CINT__) && !defined(__MAKECINT__) && !defined(__REFLEX__)
template <typename pruner>
void pruneUsing(pruner prune) {
// remove_if+erase algo applied to both vectors...
auto iter = std::find_if_not(rechits_.begin(), rechits_.end(), prune);
if (iter == rechits_.end())
return;
auto first = iter - rechits_.begin();
for (auto i = first; ++i < int(rechits_.size());) {
if (prune(rechits_[i])) {
rechits_[first] = std::move(rechits_[i]);
hitsAndFractions_[first] = std::move(hitsAndFractions_[i]);
++first;
}
}
rechits_.erase(rechits_.begin() + first, rechits_.end());
hitsAndFractions_.erase(hitsAndFractions_.begin() + first, hitsAndFractions_.end());
}
#endif
private:
/// vector of rechit fractions (transient)
std::vector<reco::PFRecHitFraction> rechits_;
/// cluster position: rho, eta, phi (transient)
REPPoint posrep_;
/// Michalis: add timing and depth information
float time_, timeError_;
double depth_;
/// transient layer
PFLayer::Layer layer_;
/// depth corrections
static const constexpr double depthCorA_ = 0.89;
static const constexpr double depthCorB_ = 7.3;
static const constexpr double depthCorAp_ = 0.89;
static const constexpr double depthCorBp_ = 4.0;
static const math::XYZPoint dummyVtx_;
};
std::ostream& operator<<(std::ostream& out, const PFCluster& cluster);
} // namespace reco
#endif // DataFormats_ParticleFlowReco_PFCluster_h
| 2,051 |
453 | #include "util.h"
#include <std/std.h>
#include <std/memory.h>
#include <kernel/util/paging/paging.h>
#include <kernel/vmm/vmm.h>
#include <kernel/boot_info.h>
void move_stack(void* new_stack_start, uint32_t size) {
Deprecated();
/*
//allocate space for new stack
printf("move_stack() mapping region 0x%08x to 0x%08x\n", new_stack_start - size, new_stack_start);
//alloc 1 extra page at the top of the stack so if you don't page fault if you
//access the very top of the stack
vmm_map_region(vmm_active_pdir(), (uint32_t*)(new_stack_start - size), size + PAGING_PAGE_SIZE, PAGE_PRESENT_FLAG|PAGE_WRITE_FLAG);
//flush TLB by reading and writing page directory address again
printf_dbg("flushing TLB");
uint32_t pd_addr;
asm volatile("mov %%cr3, %0" : "=r" (pd_addr));
asm volatile("mov %0, %%cr3" : : "r" (pd_addr));
//old ESP and EBP
uint32_t old_sp;
asm volatile("mov %%esp, %0" : "=r" (old_sp));
uint32_t old_bp;
asm volatile("mov %%ebp, %0" : "=r" (old_bp));
//offset to add to old stack addresses to get new stack address
uint32_t boot_esp = boot_info_get()->boot_stack_top_phys;
uint32_t offset = (uint32_t)new_stack_start - boot_esp;
//new esp and ebp
uint32_t new_sp = old_sp + offset;
uint32_t new_bp = old_bp + offset;
printf_dbg("new ESP %x, new EBP %x", new_sp, new_bp);
//copy stack!
printf("copying stack data!\n");
memcpy((void*)new_sp, (void*)old_sp, boot_esp - old_sp);
//backtrace through original stack, copying new values into new stack
for (uint32_t i = (uint32_t)new_sp; i > (uint32_t)new_sp - size; i -= sizeof(uint32_t)) {
uint32_t tmp = *(uint32_t*)i;
//if value of tmp is inside range of old stack,
//assume it's a base pointer and remap it
//TODO keep in mind this will remap ANY value in this range,
//whether it's a base pointer or not
if ((old_sp < tmp) && (tmp < boot_esp)) {
tmp = tmp + offset;
uint32_t* tmp2 = (uint32_t*)i;
*tmp2 = tmp;
}
}
//change stacks
asm volatile("mov %0, %%esp" : : "r" (new_sp));
asm volatile("mov %0, %%ebp" : : "r" (new_bp));
*/
}
| 845 |
440 | /*========================== begin_copyright_notice ============================
Copyright (C) 2020-2021 Intel Corporation
SPDX-License-Identifier: MIT
============================= end_copyright_notice ===========================*/
#include "wa_def.h"
#include "iadls_rev_id.h"
#define ADLS_PCH_A0_REV_ID 0
//******************* Main Wa Initializer for Device Id ********************
// Initialize COMMON/DESKTOP/MOBILE WA using PLATFORM_STEP_APPLICABLE() macro.
void InitAdlsSwWaTable(PWA_TABLE pWaTable, PSKU_FEATURE_TABLE pSkuTable, PWA_INIT_PARAM pWaParam)
{
int StepId_ADLS = (int)pWaParam->usRevId;
int PchStepId_Adls = (int)pWaParam->usRevId_PCH;
#ifdef __KCH
// compilation issue with UTF: KCHASSERT(NULL != pWaParam);
#endif
//=================================================================================================================
//
// ADLS SW WA for all platforms
//
//=================================================================================================================
SI_WA_ENABLE(
WaMixModeSelInstDstNotPacked,
"No HWBugLink provided",
"No HWSightingLink provided",
PLATFORM_ALL,
SI_WA_FOR_EVER);
}
#ifdef __KCH
void InitAdlsHASWaTable(PHW_DEVICE_EXTENSION pKchContext, PWA_TABLE pWaTable, PSKU_FEATURE_TABLE pSkuTable, PWA_INIT_PARAM pWaParam)
{
}
#endif // __KCH
| 487 |
32,544 | package com.baeldung.spring.patterns.proxy;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.stereotype.Service;
import org.springframework.transaction.annotation.Transactional;
@Service
public class BookManager {
@Autowired
private BookRepository repository;
@Transactional
public Book create(String author) {
System.out.println(repository.getClass().getName());
return repository.create(author);
}
}
| 165 |
1,738 | <gh_stars>1000+
#! /usr/bin/env python
# encoding: utf-8
# harald at klimachs.de
import re
from waflib import Utils,Errors
from waflib.Tools import fc,fc_config,fc_scan
from waflib.Configure import conf
from waflib.Tools.compiler_fc import fc_compiler
fc_compiler['aix'].insert(0, 'fc_xlf')
@conf
def find_xlf(conf):
"""Find the xlf program (will look in the environment variable 'FC')"""
fc = conf.find_program(['xlf2003_r', 'xlf2003', 'xlf95_r', 'xlf95', 'xlf90_r', 'xlf90', 'xlf_r', 'xlf'], var='FC')
fc = conf.cmd_to_list(fc)
conf.get_xlf_version(fc)
conf.env.FC_NAME='XLF'
@conf
def xlf_flags(conf):
v = conf.env
v['FCDEFINES_ST'] = '-WF,-D%s'
v['FCFLAGS_fcshlib'] = ['-qpic=small']
v['FCFLAGS_DEBUG'] = ['-qhalt=w']
v['LINKFLAGS_fcshlib'] = ['-Wl,-shared']
@conf
def xlf_modifier_platform(conf):
dest_os = conf.env['DEST_OS'] or Utils.unversioned_sys_platform()
xlf_modifier_func = getattr(conf, 'xlf_modifier_' + dest_os, None)
if xlf_modifier_func:
xlf_modifier_func()
@conf
def get_xlf_version(conf, fc):
"""Get the compiler version"""
cmd = fc + ['-qversion']
try:
out, err = conf.cmd_and_log(cmd, output=0)
except Errors.WafError:
conf.fatal('Could not find xlf %r' % cmd)
for v in (r"IBM XL Fortran.* V(?P<major>\d*)\.(?P<minor>\d*)",):
version_re = re.compile(v, re.I).search
match = version_re(out or err)
if match:
k = match.groupdict()
conf.env['FC_VERSION'] = (k['major'], k['minor'])
break
else:
conf.fatal('Could not determine the XLF version.')
def configure(conf):
conf.find_xlf()
conf.find_ar()
conf.fc_flags()
conf.fc_add_flags()
conf.xlf_flags()
conf.xlf_modifier_platform()
| 738 |
4,303 | <reponame>OAID/Halide<gh_stars>1000+
#include "runtime_internal.h"
extern "C" int sched_yield();
extern "C" WEAK void halide_thread_yield() {
sched_yield();
}
| 71 |
936 | /*
* Copyright 2020, Yahoo Inc.
* Licensed under the Apache License, Version 2.0
* See LICENSE file in project root for terms.
*/
package com.yahoo.elide.datastores.aggregation.annotation.metricformula;
import static com.yahoo.elide.datastores.aggregation.annotation.dimensionformula.DimensionFormulaTest.DUMMY_CONNECTION;
import static org.junit.jupiter.api.Assertions.assertThrows;
import static org.junit.jupiter.api.Assertions.assertTrue;
import com.yahoo.elide.core.type.ClassType;
import com.yahoo.elide.core.utils.DefaultClassScanner;
import com.yahoo.elide.datastores.aggregation.metadata.MetaDataStore;
import com.yahoo.elide.datastores.aggregation.queryengines.sql.SQLQueryEngine;
import com.google.common.collect.Sets;
import org.junit.jupiter.api.Test;
public class MetricFormulaTest {
@Test
public void testReferenceLoop() {
MetaDataStore metaDataStore = new MetaDataStore(DefaultClassScanner.getInstance(),
Sets.newHashSet(ClassType.of(MeasureLoop.class)), true);
IllegalArgumentException exception = assertThrows(
IllegalArgumentException.class,
() -> new SQLQueryEngine(metaDataStore, (unused) -> DUMMY_CONNECTION));
assertTrue(exception.getMessage().startsWith("Formula reference loop found:"));
}
}
| 455 |
1,333 | <gh_stars>1000+
/*
* Copyright 2014-2018 the original author or authors.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package de.codecentric.boot.admin.server.services.endpoints;
import de.codecentric.boot.admin.server.domain.entities.Instance;
import de.codecentric.boot.admin.server.domain.values.Endpoints;
import de.codecentric.boot.admin.server.domain.values.InstanceId;
import reactor.core.publisher.Mono;
import reactor.test.StepVerifier;
import org.junit.Test;
import static org.assertj.core.api.Assertions.assertThatThrownBy;
public class ChainingStrategyTest {
@Test
public void invariants() {
assertThatThrownBy(() -> {
new ChainingStrategy((EndpointDetectionStrategy[]) null);
}).isInstanceOf(IllegalArgumentException.class).hasMessage("'delegates' must not be null.");
assertThatThrownBy(() -> {
new ChainingStrategy((EndpointDetectionStrategy) null);
}).isInstanceOf(IllegalArgumentException.class).hasMessage("'delegates' must not contain null.");
}
@Test
public void should_chain_on_empty() {
//given
Instance instance = Instance.create(InstanceId.of("id"));
ChainingStrategy strategy = new ChainingStrategy((a) -> Mono.empty(), (a) -> Mono.empty(),
(a) -> Mono.just(Endpoints.single("id", "path")));
//when/then
StepVerifier.create(strategy.detectEndpoints(instance))
.expectNext(Endpoints.single("id", "path"))
.verifyComplete();
}
@Test
public void should_return_empty_endpoints_when_all_empty() {
//given
Instance instance = Instance.create(InstanceId.of("id"));
ChainingStrategy strategy = new ChainingStrategy((a) -> Mono.empty());
//when/then
StepVerifier.create(strategy.detectEndpoints(instance)).expectNext(Endpoints.empty()).verifyComplete();
}
}
| 865 |
438 | import numpy as np
import matplotlib as mpl
import matplotlib.pyplot as plt
import seaborn as sns
np.random.seed(sum(map(ord, "aesthetics")))
# sns.set_style("white")
x = [650, 1625, 3250, 4875, 6500]
y = [38.04, 50.28, 54.76, 56.72, 61.20]
y2 = [20.92, 31.08, 35.70, 40.16, 41.76]
plt.plot(x, y, marker='o', markersize=12, linewidth=4)
plt.plot(x, y2, marker='o', markersize=12, linewidth=4)
plt.xlabel('Number of Training Examples', fontsize=24)
plt.ylabel('Model Accuracy (%)', fontsize=26)
plt.title('DeepRegex Performance vs. Data Size', fontsize=32)
plt.legend(['DFA-Equal', 'String-Equal'], loc='upper left')
leg = plt.gca().get_legend()
ltext = leg.get_texts()
plt.setp(ltext, fontsize=18)
plt.tick_params(labelsize=16)
plt.tick_params(labelsize=16)
plt.show() | 339 |
1,041 | package org.tests.model.history;
import org.tests.model.draftable.BaseDomain;
import javax.persistence.Entity;
import javax.persistence.ManyToMany;
import java.util.List;
@Entity
public class HeDoc extends BaseDomain {
String name;
@ManyToMany(mappedBy = "docs")
List<HeLink> links;
public HeDoc(String name) {
this.name = name;
}
public String getName() {
return name;
}
public void setName(String name) {
this.name = name;
}
public List<HeLink> getLinks() {
return links;
}
public void setLinks(List<HeLink> links) {
this.links = links;
}
}
| 216 |
1,248 | <gh_stars>1000+
# -*- coding: utf-8 -*-
# Generated by Django 1.9.2 on 2016-04-23 09:44
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('pretixbase', '0020_auto_20160421_1943'),
('pretixbase', '0021_auto_20160418_2117'),
]
operations = [
]
| 151 |
1,738 | <filename>dev/Gems/Multiplayer/Code/Source/MultiplayerCVars.h
/*
* All or portions of this file Copyright (c) Amazon.com, Inc. or its affiliates or
* its licensors.
*
* For complete copyright and license terms please see the LICENSE at the root of this
* distribution (the "License"). All use of this software is governed by the License,
* or, if provided, by the license below or the license accompanying this file. Do not
* remove or modify any license notices. This file is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
*
*/
#ifndef INCLUDE_MULTIPLAYERCVARS_HEADER
#define INCLUDE_MULTIPLAYERCVARS_HEADER
#include <MultiplayerGameLiftClient.h>
#include <INetwork.h>
#include <CertificateManager/ICertificateManagerGem.h>
#include <GridMate/Session/Session.h>
struct IConsole;
struct IConsoleCmdArgs;
namespace GridMate
{
class GridSearch;
}
namespace CertificateManager
{
class FileDataSource;
}
namespace Multiplayer
{
/*!
* GridMate-specific network cvars.
*/
class MultiplayerCVars
: public GridMate::SessionEventBus::Handler
{
public:
MultiplayerCVars();
~MultiplayerCVars();
void RegisterCVars();
void UnregisterCVars();
void PostInitRegistration();
private:
//! Host a session (LAN).
static void MPHostLANCmd(IConsoleCmdArgs* args);
//! Attempt to join an existing session (LAN).
static void MPJoinLANCmd(IConsoleCmdArgs* args);
#if defined(BUILD_GAMELIFT_CLIENT)
//! Attempt to host a session on GameLift and join it.
static void MPHostGameLiftCmd(IConsoleCmdArgs* args);
//! Attempt to join an existing GameLift session.
static void MPJoinGameLiftCmd(IConsoleCmdArgs* args);
static void StopGameLiftClient(IConsoleCmdArgs* args);
static void MPMatchmakingGameLiftCmd(IConsoleCmdArgs* args);
#endif
//! Shut down current server or client session.
static void MPDisconnectCmd(IConsoleCmdArgs* args);
private:
void OnGridSearchComplete(GridMate::GridSearch* gridSearch) override;
#if defined(NET_SUPPORT_SECURE_SOCKET_DRIVER)
static void OnPrivateKeyChanged(ICVar* cvar);
static void OnCertificateChanged(ICVar* cvar);
static void OnCAChanged(ICVar* cvar);
static void CreateFileDataSource();
#endif
bool m_autoJoin;
GridMate::GridSearch* m_search;
#if defined(BUILD_GAMELIFT_CLIENT)
MultiplayerGameLiftClient m_gameLift;
#endif
static MultiplayerCVars* s_instance;
};
} // namespace GridMate
#endif // INCLUDE_NETWORKGRIDMATECVARS_HEADER
| 991 |
2,625 | /*-------------------------------------------------------------------------
*
* cmsketch.h
* Interface for Count-Min Sketch support
*
* Copyright (c) 2018, PipelineDB, Inc.
*
*-------------------------------------------------------------------------
*/
#ifndef PIPELINE_CMSKETCH_H
#define PIPELINE_CMSKETCH_H
#include "c.h"
/*
* Unfortunately we can't do top-K with Count-Min Sketch for continuous
* queries because it will require us to store O(n) values. This isn't a
* problem in the case when Count-Min Sketch structures are not being
* merged in which case only O(1) values need to be stored.
*/
typedef struct CountMinSketch
{
uint32 vl_len_;
uint64_t count;
uint32_t d;
uint32_t w;
uint32_t table[1];
} CountMinSketch;
extern CountMinSketch *CountMinSketchCreateWithDAndW(uint32_t d, uint32_t w);
extern CountMinSketch *CountMinSketchCreateWithEpsAndP(float8 epsilon, float8 p);
extern CountMinSketch *CountMinSketchCreate(void);
extern void CountMinSketchDestroy(CountMinSketch *cms);
extern CountMinSketch *CountMinSketchCopy(CountMinSketch *cms);
extern void CountMinSketchAdd(CountMinSketch *cms, void *key, Size size, uint32_t count);
extern uint32_t CountMinSketchEstimateFrequency(CountMinSketch *cms, void *key, Size size);
extern float8 CountMinSketchEstimateNormFrequency(CountMinSketch *cms, void *key, Size size);
extern uint64_t CountMinSketchTotal(CountMinSketch *cms);
extern CountMinSketch *CountMinSketchMerge(CountMinSketch *result, CountMinSketch* incoming);
extern Size CountMinSketchSize(CountMinSketch *cms);
#endif
| 526 |
997 | <gh_stars>100-1000
#include <stdint.h>
#include <string.h>
#include "address.h"
#include "hash.h"
#include "params.h"
#include "utils.h"
#include "fips202.h"
/* For SHAKE256, there is no immediate reason to initialize at the start,
so this function is an empty operation. */
void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_initialize_hash_function(
hash_state *hash_state_seeded, // NOLINT(readability-non-const-parameter)
const unsigned char *pub_seed, const unsigned char *sk_seed) {
(void)hash_state_seeded; /* Suppress an 'unused parameter' warning. */
(void)pub_seed; /* Suppress an 'unused parameter' warning. */
(void)sk_seed; /* Suppress an 'unused parameter' warning. */
}
/* This is not necessary for SHAKE256, so we don't do anything */
void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_destroy_hash_function(
hash_state *hash_state_seeded) { // NOLINT(readability-non-const-parameter)
(void)hash_state_seeded;
}
/*
* Computes PRF(key, addr), given a secret key of PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N bytes and an address
*/
void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_prf_addr(
unsigned char *out, const unsigned char *key, const uint32_t addr[8],
const hash_state *hash_state_seeded) {
unsigned char buf[PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_ADDR_BYTES];
memcpy(buf, key, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N);
PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_addr_to_bytes(buf + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N, addr);
shake256(out, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N, buf, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_ADDR_BYTES);
(void)hash_state_seeded; /* Prevent unused parameter warning. */
}
/**
* Computes the message-dependent randomness R, using a secret seed and an
* optional randomization value as well as the message.
*/
void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_gen_message_random(
unsigned char *R,
const unsigned char *sk_prf, const unsigned char *optrand,
const unsigned char *m, size_t mlen,
const hash_state *hash_state_seeded) {
shake256incctx state;
shake256_inc_init(&state);
shake256_inc_absorb(&state, sk_prf, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N);
shake256_inc_absorb(&state, optrand, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N);
shake256_inc_absorb(&state, m, mlen);
shake256_inc_finalize(&state);
shake256_inc_squeeze(R, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N, &state);
shake256_inc_ctx_release(&state);
(void)hash_state_seeded; /* Prevent unused parameter warning. */
}
/**
* Computes the message hash using R, the public key, and the message.
* Outputs the message digest and the index of the leaf. The index is split in
* the tree index and the leaf index, for convenient copying to an address.
*/
void PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_hash_message(
unsigned char *digest, uint64_t *tree, uint32_t *leaf_idx,
const unsigned char *R, const unsigned char *pk,
const unsigned char *m, size_t mlen,
const hash_state *hash_state_seeded) {
#define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BITS (PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_HEIGHT * (PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_D - 1))
#define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BYTES ((PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BITS + 7) / 8)
#define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BITS PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_HEIGHT
#define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BYTES ((PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BITS + 7) / 8)
#define PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_DGST_BYTES (PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_FORS_MSG_BYTES + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BYTES + PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BYTES)
unsigned char buf[PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_DGST_BYTES];
unsigned char *bufp = buf;
shake256incctx state;
shake256_inc_init(&state);
shake256_inc_absorb(&state, R, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_N);
shake256_inc_absorb(&state, pk, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_PK_BYTES);
shake256_inc_absorb(&state, m, mlen);
shake256_inc_finalize(&state);
shake256_inc_squeeze(buf, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_DGST_BYTES, &state);
shake256_inc_ctx_release(&state);
memcpy(digest, bufp, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_FORS_MSG_BYTES);
bufp += PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_FORS_MSG_BYTES;
*tree = PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_bytes_to_ull(
bufp, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BYTES);
*tree &= (~(uint64_t)0) >> (64 - PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BITS);
bufp += PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_TREE_BYTES;
*leaf_idx = (uint32_t)PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_bytes_to_ull(
bufp, PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BYTES);
*leaf_idx &= (~(uint32_t)0) >> (32 - PQCLEAN_SPHINCSSHAKE256256SSIMPLE_AVX2_LEAF_BITS);
(void)hash_state_seeded; /* Prevent unused parameter warning. */
}
| 2,282 |
1,582 | # -*- coding: utf-8 -*-
import json
from vilya.libs import api_errors
from vilya.models.project import CodeDoubanProject
from vilya.views.api.utils import RestAPIUI, api_require_login, jsonize
from vilya.views.api.repos.product import ProductUI
from vilya.views.api.repos.summary import SummaryUI
from vilya.views.api.repos.intern import InternUI
from vilya.views.api.repos.default_branch import DefaultBranchUI
from vilya.views.api.repos.commits import CommitsUI
from vilya.views.api.repos.post_receive import PostReceiveUI
from vilya.views.api.repos.git2svn import GIT2SVNUI
from vilya.views.api.repos.svn2git import SVN2GITUI
from vilya.views.api.repos.pulls import PullsUI
from vilya.views.api.repos.issues import IssuesUI
from vilya.views.api.repos.contents import ContentsUI
from vilya.views.api.repos.push import PushUI
from vilya.views.api.repos.watchers import WatchersUI
_q_exports = []
def _q_lookup(request, name):
return RepositoryUI(name)
def _q_access(request):
request.response.set_content_type('application/json; charset=utf-8')
class RepositoryUI(object):
_q_exports = [
'lang_stats', 'forks', 'pulls', 'summary',
'committers', 'name', 'owner', 'product',
'intern_banned', 'default_branch', 'commits',
'post_receive', 'svn2git', 'git2svn', 'issues',
'contents', 'can_push', 'watchers'
]
def __init__(self, name):
self.name = name
self.repo = CodeDoubanProject.get_by_name(self.name)
def __call__(self, request):
return self._q_index(request)
@jsonize
def _q_index(self, request):
if not self.repo:
raise api_errors.NotFoundError("repo")
return {}
def _q_access(self, request):
self.method = request.method
def _q_lookup(self, request, part):
name = "%s/%s" % (self.name, part)
if not CodeDoubanProject.exists(name):
raise api_errors.NotFoundError("repo")
return RepositoryUI(name)
@jsonize
def lang_stats(self, request):
if not self.repo:
raise api_errors.NotFoundError
if self.method == 'POST':
language = request.get_form_var('language', '')
languages = request.get_form_var('languages', '[]')
try:
languages = json.loads(languages)
except ValueError:
raise api_errors.NotJSONError
self.repo.language = language
self.repo.languages = languages
return {}
else:
return dict(language=self.repo.language,
languages=self.repo.languages)
@property
def forks(self):
return ForksUI(self.repo)
@property
def pulls(self):
return PullsUI(self.repo)
@property
def product(self):
return ProductUI(self.repo)
@property
def summary(self):
return SummaryUI(self.repo)
@property
def intern_banned(self):
return InternUI(self.repo)
@property
def can_push(self):
return PushUI(self.repo)
@property
def default_branch(self):
return DefaultBranchUI(self.repo)
@property
def commits(self):
return CommitsUI(self.repo)
@property
def post_receive(self):
return PostReceiveUI(self.repo)
@property
def svn2git(self):
return SVN2GITUI(self.repo)
@property
def git2svn(self):
return GIT2SVNUI(self.repo)
@property
def issues(self):
return IssuesUI(self.repo)
@property
def contents(self):
return ContentsUI(self.repo)
@property
def watchers(self):
return WatchersUI(self.repo)
class ForksUI(RestAPIUI):
_q_exports = []
_q_methods = ['get', 'post']
def __init__(self, repo):
self.repo = repo
@api_require_login
def post(self, request):
repo = self.repo
fork_repo = repo.new_fork(self.user.name)
if not fork_repo:
# FIXME: repository exists
return []
return fork_repo.as_dict()
def get(self, request):
fork_repos = self.repo.get_forked_projects()
return [project.get_info(without_commits=True)
for project in fork_repos]
| 1,899 |
471 | <reponame>madanagopaltcomcast/pxCore
/////////////////////////////////////////////////////////////////////////////
// Name: wx/gtk1/checkbox.h
// Purpose:
// Author: <NAME>
// Copyright: (c) 1998 <NAME>
// Licence: wxWindows licence
/////////////////////////////////////////////////////////////////////////////
#ifndef __GTKCHECKBOXH__
#define __GTKCHECKBOXH__
// ----------------------------------------------------------------------------
// wxCheckBox
// ----------------------------------------------------------------------------
class WXDLLIMPEXP_CORE wxCheckBox : public wxCheckBoxBase
{
public:
wxCheckBox();
wxCheckBox( wxWindow *parent, wxWindowID id, const wxString& label,
const wxPoint& pos = wxDefaultPosition,
const wxSize& size = wxDefaultSize, long style = 0,
const wxValidator& validator = wxDefaultValidator,
const wxString& name = wxCheckBoxNameStr)
{
Create(parent, id, label, pos, size, style, validator, name);
}
bool Create(wxWindow *parent,
wxWindowID id,
const wxString& label,
const wxPoint& pos = wxDefaultPosition,
const wxSize& size = wxDefaultSize,
long style = 0,
const wxValidator& validator = wxDefaultValidator,
const wxString& name = wxCheckBoxNameStr );
void SetValue( bool state );
bool GetValue() const;
virtual void SetLabel( const wxString& label );
virtual bool Enable( bool enable = TRUE );
static wxVisualAttributes
GetClassDefaultAttributes(wxWindowVariant variant = wxWINDOW_VARIANT_NORMAL);
// implementation
// --------------
void DoApplyWidgetStyle(GtkRcStyle *style);
bool IsOwnGtkWindow( GdkWindow *window );
void OnInternalIdle();
GtkWidget *m_widgetCheckbox;
GtkWidget *m_widgetLabel;
bool m_blockEvent;
protected:
virtual wxSize DoGetBestSize() const;
private:
DECLARE_DYNAMIC_CLASS(wxCheckBox)
};
#endif // __GTKCHECKBOXH__
| 777 |
4,036 | <reponame>vadi2/codeql<filename>cpp/ql/test/library-tests/proxy_class/proxy_class.cpp
struct Base {};
template <typename T>
struct Derived {friend class T;};
int main() { return sizeof(Derived<Base>); }
// semmle-extractor-options: --microsoft
| 87 |
5,964 | #ifndef V8_TORQUE_TYPED_ARRAY_SUBARRAY_FROM_DSL_BASE_H__
#define V8_TORQUE_TYPED_ARRAY_SUBARRAY_FROM_DSL_BASE_H__
#include "src/compiler/code-assembler.h"
#include "src/code-stub-assembler.h"
#include "src/utils.h"
#include "torque-generated/class-definitions-from-dsl.h"
namespace v8 {
namespace internal {
class TypedArraySubarrayBuiltinsFromDSLAssembler {
public:
explicit TypedArraySubarrayBuiltinsFromDSLAssembler(compiler::CodeAssemblerState* state) : state_(state), ca_(state) { USE(state_, ca_); }
private:
compiler::CodeAssemblerState* const state_;
compiler::CodeAssembler ca_;
};
} // namespace internal
} // namespace v8
#endif // V8_TORQUE_TYPED_ARRAY_SUBARRAY_FROM_DSL_BASE_H__
| 291 |
2,293 | _G='r'
_F='w'
_E=False
_D=True
_C='strict'
_B='utf-8'
_A=None
import csv,json,sys,warnings
from pathlib import Path
import dynaconf.vendor.ruamel.yaml as yaml
from dynaconf.vendor.box.exceptions import BoxError,BoxWarning
from dynaconf.vendor import toml
BOX_PARAMETERS='default_box','default_box_attr','conversion_box','frozen_box','camel_killer_box','box_safe_prefix','box_duplicates','ordered_box','default_box_none_transform','box_dots','modify_tuples_box','box_intact_types','box_recast'
def _exists(filename,create=_E):
A=filename;B=Path(A)
if create:
try:B.touch(exist_ok=_D)
except OSError as C:raise BoxError(f"Could not create file {A} - {C}")
else:return
if not B.exists():raise BoxError(f'File "{A}" does not exist')
if not B.is_file():raise BoxError(f"{A} is not a file")
def _to_json(obj,filename=_A,encoding=_B,errors=_C,**C):
A=filename;B=json.dumps(obj,ensure_ascii=_E,**C)
if A:
_exists(A,create=_D)
with open(A,_F,encoding=encoding,errors=errors)as D:D.write(B if sys.version_info>=(3,0)else B.decode(_B))
else:return B
def _from_json(json_string=_A,filename=_A,encoding=_B,errors=_C,multiline=_E,**B):
D=json_string;A=filename
if A:
_exists(A)
with open(A,_G,encoding=encoding,errors=errors)as E:
if multiline:C=[json.loads(A.strip(),**B)for A in E if A.strip()and not A.strip().startswith('#')]
else:C=json.load(E,**B)
elif D:C=json.loads(D,**B)
else:raise BoxError('from_json requires a string or filename')
return C
def _to_yaml(obj,filename=_A,default_flow_style=_E,encoding=_B,errors=_C,**C):
B=default_flow_style;A=filename
if A:
_exists(A,create=_D)
with open(A,_F,encoding=encoding,errors=errors)as D:yaml.dump(obj,stream=D,default_flow_style=B,**C)
else:return yaml.dump(obj,default_flow_style=B,**C)
def _from_yaml(yaml_string=_A,filename=_A,encoding=_B,errors=_C,**A):
F='Loader';C=yaml_string;B=filename
if F not in A:A[F]=yaml.SafeLoader
if B:
_exists(B)
with open(B,_G,encoding=encoding,errors=errors)as E:D=yaml.load(E,**A)
elif C:D=yaml.load(C,**A)
else:raise BoxError('from_yaml requires a string or filename')
return D
def _to_toml(obj,filename=_A,encoding=_B,errors=_C):
A=filename
if A:
_exists(A,create=_D)
with open(A,_F,encoding=encoding,errors=errors)as B:toml.dump(obj,B)
else:return toml.dumps(obj)
def _from_toml(toml_string=_A,filename=_A,encoding=_B,errors=_C):
B=toml_string;A=filename
if A:
_exists(A)
with open(A,_G,encoding=encoding,errors=errors)as D:C=toml.load(D)
elif B:C=toml.loads(B)
else:raise BoxError('from_toml requires a string or filename')
return C
def _to_csv(box_list,filename,encoding=_B,errors=_C):
B=filename;A=box_list;C=list(A[0].keys())
for E in A:
if list(E.keys())!=C:raise BoxError('BoxList must contain the same dictionary structure for every item to convert to csv')
if B:
_exists(B,create=_D)
with open(B,_F,encoding=encoding,errors=errors,newline='')as F:
D=csv.DictWriter(F,fieldnames=C);D.writeheader()
for G in A:D.writerow(G)
def _from_csv(filename,encoding=_B,errors=_C):
A=filename;_exists(A)
with open(A,_G,encoding=encoding,errors=errors,newline='')as B:C=csv.DictReader(B);return[A for A in C] | 1,410 |
471 | <reponame>dimagilg/commcare-hq
class NoAccountException(Exception):
"""
Raised when trying to access the account of someone without one
"""
pass
class InvalidMobileWorkerRequest(Exception):
pass
class IllegalAccountConfirmation(Exception):
pass
| 87 |
2,231 | #!/usr/bin/env python
# Copyright 2020 The Defold Foundation
# Licensed under the Defold License version 1.0 (the "License"); you may not use
# this file except in compliance with the License.
#
# You may obtain a copy of the License, together with FAQs at
# https://www.defold.com/license
#
# Unless required by applicable law or agreed to in writing, software distributed
# under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
# CONDITIONS OF ANY KIND, either express or implied. See the License for the
# specific language governing permissions and limitations under the License.
import Tkinter
import glob, stat, os, urllib
root_path = "content"
state = {}
def UpdateState():
changed = []
for path, dirs, files in os.walk(root_path):
(root, sep, sub_path) = path.partition(os.path.sep)
for f in files:
name = os.path.join(path, f)
mtime = os.stat(name)[stat.ST_MTIME]
oldmtime = state.get(name, 0)
if mtime != oldmtime:
changed.append(os.path.join(sub_path, f + "c"))
state[name] = mtime
return changed
root = Tkinter.Tk()
root.minsize(300, 200)
label = Tkinter.Label()
label.config(text = "Waiting for resources to change")
label.pack()
UpdateState()
def tick():
label.config(text = "Waiting for resources to change")
changed = UpdateState()
if len(changed) > 0:
label.config(text = "Compiling...")
label.update()
result = os.system('waf --skip-tests')
if result == 0:
try:
for f in changed:
print("Reloading: " + f)
request = urllib.urlopen("http://localhost:8001/reload/" + f)
label.config(text = "Success!", foreground = "#080")
except:
label.config(text = "No connection!", foreground = "#880")
else:
label.config(text = "Failure!", foreground = "#a00")
label.after(1000, tick)
tick()
label.mainloop()
| 818 |
603 | """CFFI backend (for PyPy)"""
# Copyright (C) PyZMQ Developers
# Distributed under the terms of the Modified BSD License.
from zmq.backend.cffi import (
constants,
error,
message,
context,
socket,
_poll,
devices,
utils,
)
from ._cffi import ffi, lib as C
def zmq_version_info():
"""Get libzmq version as tuple of ints"""
major = ffi.new('int*')
minor = ffi.new('int*')
patch = ffi.new('int*')
C.zmq_version(major, minor, patch)
return (int(major[0]), int(minor[0]), int(patch[0]))
__all__ = ["zmq_version_info"]
for submod in (constants, error, message, context, socket, _poll, devices, utils):
__all__.extend(submod.__all__)
from .constants import *
from .error import *
from .message import *
from .context import *
from .socket import *
from .devices import *
from ._poll import *
from .utils import *
| 339 |
2,114 | package io.searchbox.client.http;
import com.google.gson.Gson;
import io.searchbox.action.Action;
import io.searchbox.client.AbstractJestClient;
import io.searchbox.client.JestResult;
import io.searchbox.client.JestResultHandler;
import io.searchbox.client.config.ElasticsearchVersion;
import io.searchbox.client.config.exception.CouldNotConnectException;
import io.searchbox.client.http.apache.HttpDeleteWithEntity;
import io.searchbox.client.http.apache.HttpGetWithEntity;
import org.apache.http.Header;
import org.apache.http.HttpEntityEnclosingRequest;
import org.apache.http.HttpRequest;
import org.apache.http.HttpResponse;
import org.apache.http.StatusLine;
import org.apache.http.client.config.RequestConfig;
import org.apache.http.client.entity.EntityBuilder;
import org.apache.http.client.methods.CloseableHttpResponse;
import org.apache.http.client.methods.HttpHead;
import org.apache.http.client.methods.HttpPost;
import org.apache.http.client.methods.HttpPut;
import org.apache.http.client.methods.HttpRequestBase;
import org.apache.http.client.methods.HttpUriRequest;
import org.apache.http.client.protocol.HttpClientContext;
import org.apache.http.concurrent.FutureCallback;
import org.apache.http.conn.HttpHostConnectException;
import org.apache.http.entity.ContentType;
import org.apache.http.impl.client.CloseableHttpClient;
import org.apache.http.impl.nio.client.CloseableHttpAsyncClient;
import org.apache.http.util.EntityUtils;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.io.IOException;
import java.util.Map.Entry;
import java.util.concurrent.Future;
/**
* @author <NAME>
* @author <NAME>
*/
public class JestHttpClient extends AbstractJestClient {
private final static Logger log = LoggerFactory.getLogger(JestHttpClient.class);
protected ContentType requestContentType = ContentType.APPLICATION_JSON.withCharset("utf-8");
private CloseableHttpClient httpClient;
private CloseableHttpAsyncClient asyncClient;
private HttpClientContext httpClientContextTemplate;
private ElasticsearchVersion elasticsearchVersion = ElasticsearchVersion.UNKNOWN;
/**
* @throws IOException in case of a problem or the connection was aborted during request,
* or in case of a problem while reading the response stream
* @throws CouldNotConnectException if an {@link HttpHostConnectException} is encountered
*/
@Override
public <T extends JestResult> T execute(Action<T> clientRequest) throws IOException {
return execute(clientRequest, null);
}
public <T extends JestResult> T execute(Action<T> clientRequest, RequestConfig requestConfig) throws IOException {
HttpUriRequest request = prepareRequest(clientRequest, requestConfig);
CloseableHttpResponse response = null;
try {
response = executeRequest(request);
return deserializeResponse(response, request, clientRequest);
} catch (HttpHostConnectException ex) {
throw new CouldNotConnectException(ex.getHost().toURI(), ex);
} finally {
if (response != null) {
try {
response.close();
} catch (IOException ex) {
log.error("Exception occurred while closing response stream.", ex);
}
}
}
}
@Override
public <T extends JestResult> void executeAsync(final Action<T> clientRequest, final JestResultHandler<? super T> resultHandler) {
executeAsync(clientRequest, resultHandler, null);
}
public <T extends JestResult> void executeAsync(final Action<T> clientRequest, final JestResultHandler<? super T> resultHandler, final RequestConfig requestConfig) {
synchronized (this) {
if (!asyncClient.isRunning()) {
asyncClient.start();
}
}
HttpUriRequest request = prepareRequest(clientRequest, requestConfig);
executeAsyncRequest(clientRequest, resultHandler, request);
}
@Override
public void shutdownClient() {
try {
close();
} catch (IOException e) {
log.error("Exception occurred while shutting down the sync client.", e);
}
}
@Override
public void close() throws IOException {
super.close();
asyncClient.close();
httpClient.close();
}
protected <T extends JestResult> HttpUriRequest prepareRequest(final Action<T> clientRequest, final RequestConfig requestConfig) {
String elasticSearchRestUrl = getRequestURL(getNextServer(), clientRequest.getURI(elasticsearchVersion));
HttpUriRequest request = constructHttpMethod(clientRequest.getRestMethodName(), elasticSearchRestUrl, clientRequest.getData(gson), requestConfig);
log.debug("Request method={} url={}", clientRequest.getRestMethodName(), elasticSearchRestUrl);
// add headers added to action
for (Entry<String, Object> header : clientRequest.getHeaders().entrySet()) {
request.addHeader(header.getKey(), header.getValue().toString());
}
return request;
}
protected CloseableHttpResponse executeRequest(HttpUriRequest request) throws IOException {
if (httpClientContextTemplate != null) {
return httpClient.execute(request, createContextInstance());
}
return httpClient.execute(request);
}
protected <T extends JestResult> Future<HttpResponse> executeAsyncRequest(Action<T> clientRequest, JestResultHandler<? super T> resultHandler, HttpUriRequest request) {
if (httpClientContextTemplate != null) {
return asyncClient.execute(request, createContextInstance(), new DefaultCallback<T>(clientRequest, request, resultHandler));
}
return asyncClient.execute(request, new DefaultCallback<T>(clientRequest, request, resultHandler));
}
protected HttpClientContext createContextInstance() {
HttpClientContext context = HttpClientContext.create();
context.setCredentialsProvider(httpClientContextTemplate.getCredentialsProvider());
context.setAuthCache(httpClientContextTemplate.getAuthCache());
return context;
}
protected HttpUriRequest constructHttpMethod(String methodName, String url, String payload, RequestConfig requestConfig) {
HttpUriRequest httpUriRequest = null;
if (methodName.equalsIgnoreCase("POST")) {
httpUriRequest = new HttpPost(url);
log.debug("POST method created based on client request");
} else if (methodName.equalsIgnoreCase("PUT")) {
httpUriRequest = new HttpPut(url);
log.debug("PUT method created based on client request");
} else if (methodName.equalsIgnoreCase("DELETE")) {
httpUriRequest = new HttpDeleteWithEntity(url);
log.debug("DELETE method created based on client request");
} else if (methodName.equalsIgnoreCase("GET")) {
httpUriRequest = new HttpGetWithEntity(url);
log.debug("GET method created based on client request");
} else if (methodName.equalsIgnoreCase("HEAD")) {
httpUriRequest = new HttpHead(url);
log.debug("HEAD method created based on client request");
}
if (httpUriRequest instanceof HttpRequestBase && requestConfig != null) {
((HttpRequestBase) httpUriRequest).setConfig(requestConfig);
}
if (httpUriRequest != null && httpUriRequest instanceof HttpEntityEnclosingRequest && payload != null) {
EntityBuilder entityBuilder = EntityBuilder.create()
.setText(payload)
.setContentType(requestContentType);
if (isRequestCompressionEnabled()) {
entityBuilder.gzipCompress();
}
((HttpEntityEnclosingRequest) httpUriRequest).setEntity(entityBuilder.build());
}
return httpUriRequest;
}
private <T extends JestResult> T deserializeResponse(HttpResponse response, final HttpRequest httpRequest, Action<T> clientRequest) throws IOException {
StatusLine statusLine = response.getStatusLine();
try {
return clientRequest.createNewElasticSearchResult(
response.getEntity() == null ? null : EntityUtils.toString(response.getEntity()),
statusLine.getStatusCode(),
statusLine.getReasonPhrase(),
gson
);
} catch (com.google.gson.JsonSyntaxException e) {
for (Header header : response.getHeaders("Content-Type")) {
final String mimeType = header.getValue();
if (!mimeType.startsWith("application/json")) {
// probably a proxy that responded in text/html
final String message = "Request " + httpRequest.toString() + " yielded " + mimeType
+ ", should be json: " + statusLine.toString();
throw new IOException(message, e);
}
}
throw e;
}
}
public CloseableHttpClient getHttpClient() {
return httpClient;
}
public void setHttpClient(CloseableHttpClient httpClient) {
this.httpClient = httpClient;
}
public CloseableHttpAsyncClient getAsyncClient() {
return asyncClient;
}
public void setAsyncClient(CloseableHttpAsyncClient asyncClient) {
this.asyncClient = asyncClient;
}
public Gson getGson() {
return gson;
}
public void setGson(Gson gson) {
this.gson = gson;
}
public HttpClientContext getHttpClientContextTemplate() {
return httpClientContextTemplate;
}
public void setHttpClientContextTemplate(HttpClientContext httpClientContext) {
this.httpClientContextTemplate = httpClientContext;
}
public void setElasticsearchVersion(ElasticsearchVersion elasticsearchVersion) {
this.elasticsearchVersion = elasticsearchVersion;
}
protected class DefaultCallback<T extends JestResult> implements FutureCallback<HttpResponse> {
private final Action<T> clientRequest;
private final HttpRequest request;
private final JestResultHandler<? super T> resultHandler;
public DefaultCallback(Action<T> clientRequest, final HttpRequest request, JestResultHandler<? super T> resultHandler) {
this.clientRequest = clientRequest;
this.request = request;
this.resultHandler = resultHandler;
}
@Override
public void completed(final HttpResponse response) {
T jestResult = null;
try {
jestResult = deserializeResponse(response, request, clientRequest);
} catch (Exception e) {
failed(e);
} catch (Throwable t) {
failed(new Exception("Problem during request processing", t));
}
if (jestResult != null) resultHandler.completed(jestResult);
}
@Override
public void failed(final Exception ex) {
log.error("Exception occurred during async execution.", ex);
if (ex instanceof HttpHostConnectException) {
String host = ((HttpHostConnectException) ex).getHost().toURI();
resultHandler.failed(new CouldNotConnectException(host, ex));
return;
}
resultHandler.failed(ex);
}
@Override
public void cancelled() {
log.warn("Async execution was cancelled; this is not expected to occur under normal operation.");
}
}
}
| 4,401 |
679 | <reponame>Grosskopf/openoffice
/**************************************************************
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*
*************************************************************/
#ifndef SC_EDITSRC_HXX
#define SC_EDITSRC_HXX
#include "address.hxx"
#include <editeng/unoedsrc.hxx>
#include <svl/lstner.hxx>
#include <memory>
class ScEditEngineDefaulter;
class SvxEditEngineForwarder;
class ScDocShell;
class ScHeaderFooterContentObj;
class ScCellTextData;
class ScHeaderFooterTextData;
class ScAccessibleTextData;
class SdrObject;
class ScHeaderFooterChangedHint : public SfxHint
{
sal_uInt16 nPart;
public:
TYPEINFO();
ScHeaderFooterChangedHint(sal_uInt16 nP);
~ScHeaderFooterChangedHint();
sal_uInt16 GetPart() const { return nPart; }
};
// all ScSharedHeaderFooterEditSource objects for a single text share the same data
class ScSharedHeaderFooterEditSource : public SvxEditSource
{
private:
ScHeaderFooterTextData* pTextData;
protected:
ScHeaderFooterTextData* GetTextData() const { return pTextData; } // for ScHeaderFooterEditSource
public:
ScSharedHeaderFooterEditSource( ScHeaderFooterTextData* pData );
virtual ~ScSharedHeaderFooterEditSource();
// GetEditEngine is needed because the forwarder doesn't have field functions
ScEditEngineDefaulter* GetEditEngine();
virtual SvxEditSource* Clone() const ;
virtual SvxTextForwarder* GetTextForwarder();
virtual void UpdateData();
};
// ScHeaderFooterEditSource with local copy of ScHeaderFooterTextData is used by field objects
class ScHeaderFooterEditSource : public ScSharedHeaderFooterEditSource
{
public:
ScHeaderFooterEditSource( ScHeaderFooterContentObj* pContent, sal_uInt16 nP );
ScHeaderFooterEditSource( ScHeaderFooterContentObj& rContent, sal_uInt16 nP );
virtual ~ScHeaderFooterEditSource();
virtual SvxEditSource* Clone() const;
};
// Data (incl. EditEngine) for cell EditSource is now shared in ScCellTextData
class ScSharedCellEditSource : public SvxEditSource
{
private:
ScCellTextData* pCellTextData;
protected:
ScCellTextData* GetCellTextData() const { return pCellTextData; } // for ScCellEditSource
public:
ScSharedCellEditSource( ScCellTextData* pData );
virtual ~ScSharedCellEditSource();
// GetEditEngine is needed because the forwarder doesn't have field functions
ScEditEngineDefaulter* GetEditEngine();
virtual SvxEditSource* Clone() const;
virtual SvxTextForwarder* GetTextForwarder();
virtual void UpdateData();
void SetDoUpdateData(sal_Bool bValue);
sal_Bool IsDirty() const;
};
// ScCellEditSource with local copy of ScCellTextData is used by ScCellFieldsObj, ScCellFieldObj
class ScCellEditSource : public ScSharedCellEditSource
{
public:
ScCellEditSource( ScDocShell* pDocSh, const ScAddress& rP );
virtual ~ScCellEditSource();
virtual SvxEditSource* Clone() const;
};
class ScAnnotationEditSource : public SvxEditSource, public SfxListener
{
private:
ScDocShell* pDocShell;
ScAddress aCellPos;
ScEditEngineDefaulter* pEditEngine;
SvxEditEngineForwarder* pForwarder;
sal_Bool bDataValid;
SdrObject* GetCaptionObj();
public:
ScAnnotationEditSource(ScDocShell* pDocSh, const ScAddress& rP);
virtual ~ScAnnotationEditSource();
virtual SvxEditSource* Clone() const ;
virtual SvxTextForwarder* GetTextForwarder();
virtual void UpdateData();
virtual void Notify( SfxBroadcaster& rBC, const SfxHint& rHint );
};
// EditSource with a shared forwarder for all children of one text object
class ScSimpleEditSource : public SvxEditSource
{
private:
SvxTextForwarder* pForwarder;
public:
ScSimpleEditSource( SvxTextForwarder* pForw );
virtual ~ScSimpleEditSource();
virtual SvxEditSource* Clone() const ;
virtual SvxTextForwarder* GetTextForwarder();
virtual void UpdateData();
};
class ScAccessibilityEditSource : public SvxEditSource
{
private:
::std::auto_ptr < ScAccessibleTextData > mpAccessibleTextData;
public:
ScAccessibilityEditSource( ::std::auto_ptr < ScAccessibleTextData > pAccessibleCellTextData );
virtual ~ScAccessibilityEditSource();
virtual SvxEditSource* Clone() const;
virtual SvxTextForwarder* GetTextForwarder();
virtual SvxViewForwarder* GetViewForwarder();
virtual SvxEditViewForwarder* GetEditViewForwarder( sal_Bool bCreate = sal_False );
virtual void UpdateData();
virtual SfxBroadcaster& GetBroadcaster() const;
};
#endif
| 1,778 |
17,481 | /*
* Copyright (C) 2021 The Dagger Authors.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package dagger.hilt.android.internal;
import android.app.Application;
import android.content.Context;
import android.content.ContextWrapper;
/**
* Utility methods for dealing with contexts.
*/
public final class Contexts {
/** Finds the android Application from a context. */
public static Application getApplication(Context context) {
if (context instanceof Application) {
return (Application) context;
}
Context unwrapContext = context;
while (unwrapContext instanceof ContextWrapper) {
unwrapContext = ((ContextWrapper) unwrapContext).getBaseContext();
if (unwrapContext instanceof Application) {
return (Application) unwrapContext;
}
}
throw new IllegalStateException(
"Could not find an Application in the given context: " + context);
}
private Contexts() {}
}
| 406 |
10,225 | package org.jboss.resteasy.reactive.common.providers.serialisers;
import java.io.ByteArrayOutputStream;
import java.io.IOException;
import java.io.InputStream;
import java.nio.charset.StandardCharsets;
import javax.ws.rs.core.MediaType;
public class MessageReaderUtil {
public static final String UTF8_CHARSET = StandardCharsets.UTF_8.name();
public static String charsetFromMediaType(MediaType mediaType) {
if (mediaType == null) {
return UTF8_CHARSET;
}
String charset = mediaType.getParameters().get(MediaType.CHARSET_PARAMETER);
if (charset != null) {
return charset;
}
return UTF8_CHARSET;
}
public static byte[] readBytes(InputStream entityStream) throws IOException {
ByteArrayOutputStream out = new ByteArrayOutputStream();
byte[] buf = new byte[1024]; //TODO: fix, needs a pure vert.x async read model
int r;
while ((r = entityStream.read(buf)) > 0) {
out.write(buf, 0, r);
}
return out.toByteArray();
}
public static String readString(InputStream entityStream, MediaType mediaType) throws IOException {
return new String(readBytes(entityStream), charsetFromMediaType(mediaType));
}
}
| 492 |
310 | <gh_stars>100-1000
{
"name": "import.io",
"description": "A service for scraping data from websites.",
"url": "https://www.import.io/"
} | 50 |
2,462 | package com.sanshengshui.persistence;
import com.sanshengshui.persistence.entity.ToData;
import java.util.*;
public abstract class DaoUtil {
private DaoUtil() {
}
public static <T> List<T> convertDataList(Collection<? extends ToData<T>> toDataList) {
List<T> list = Collections.emptyList();
if (toDataList != null && !toDataList.isEmpty()) {
list = new ArrayList<>();
for (ToData<T> object : toDataList) {
if (object != null) {
list.add(object.toData());
}
}
}
return list;
}
public static <T> T getData(ToData<T> data) {
T object = null;
if (data != null) {
object = data.toData();
}
return object;
}
public static Long getId(Long idBased) {
Long id = null;
if (idBased != null) {
id = idBased;
}
return id;
}
public static List<Long> toUUIDs(List<? extends Long> idBasedIds) {
List<Long> ids = new ArrayList<>();
for (Long idBased : idBasedIds) {
ids.add(getId(idBased));
}
return ids;
}
}
| 591 |
679 | <gh_stars>100-1000
/**************************************************************
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*
*************************************************************/
package org.openoffice.xmerge.converter.xml;
import org.w3c.dom.Node;
import org.w3c.dom.NamedNodeMap;
/**
* An object of class <code>Style</code> represents a <i>style</i>
* in an OpenOffice document. In practice subclasses of this
* <code>Style</code>, such as <code>TextStyle</code>,
* <code>ParaStyle</code> are used.
*
* @author <NAME>
* @see <a href="TextStyle.html">TextStyle</a>,
* <a href="ParaStyle.html">ParaStyle</a>
*/
public class Style {
/** Name of the <code>Style</code>. */
protected String name = null;
/** Family of the <code>Style</code>. */
protected String family = null;
/** Parent of the <code>Style</code>. */
protected String parent = null;
/**
* A reference to the <code>StyleCatalog</code> to be used for
* looking up ancestor <code>Style</code> objects.
*/
protected StyleCatalog sc;
/**
* Constructor for use when going from DOM to client device format.
*
* @param node A <i>style:style</i> or <i>style:default-style</i>
* <code>Node</code> from the document being parsed.
* No checking of <code>Node</code> is done, so if it
* is not of the proper type the results will be
* unpredictable.
* @param sc The <code>StyleCatalog</code>, which is used for
* looking up ancestor <code>Style</code> objects.
*/
public Style(Node node, StyleCatalog sc) {
this.sc = sc;
// Run through the attributes of this node, saving
// the ones we're interested in.
if (node.getNodeName().equals("style:default-style"))
name = "DEFAULT_STYLE";
NamedNodeMap attrNodes = node.getAttributes();
if (attrNodes != null) {
int len = attrNodes.getLength();
for (int i = 0; i < len; i++) {
Node attr = attrNodes.item(i);
if (attr.getNodeName().equals("style:family"))
family = attr.getNodeValue();
else if (attr.getNodeName().equals("style:name")) {
name = attr.getNodeValue();
} else if (attr.getNodeName().equals("style:parent-style-name"))
parent = attr.getNodeValue();
}
}
}
/**
* Constructor for use when going from client device format to DOM.
*
* @param name Name of the <code>Style</code>. Can be null.
* @param family Family of the <code>Style</code> - usually
* <i>paragraph</i>, <i>text</i>, etc. Can be null.
* @param parent Name of the parent <code>Style</code>, or null if none.
* @param sc The <code>StyleCatalog</code>, which is used for
* looking up ancestor <code>Style</code> objects.
*/
public Style(String name, String family, String parent, StyleCatalog sc) {
this.sc = sc;
this.name = name;
this.family = family;
this.parent = parent;
}
/**
* Set the <code>StyleCatalog</code> to be used when looking up the
* <code>Style</code> parent.
*
* @param sc The <code>StyleCatalog</code>, which is used for
* looking up ancestor <code>Style</code> objects.
*/
public void setCatalog(StyleCatalog sc) {
this.sc = sc;
}
/**
* Returns the name of this <code>Style</code>.
*
* @return The name of this <code>Style</code>.
*/
public String getName() {
return name;
}
/**
* Sets the name of this <code>Style</code>.
*
* @param newName The new name of this <code>Style</code>.
*/
public void setName(String newName) {
name = newName;
}
/**
* Return the family of this <code>Style</code>.
*
* @return The family of this <code>Style</code>.
*/
public String getFamily() {
return family;
}
/**
* Return the name of the parent of this <code>Style</code>.
*
* @return The parent of this <code>Style</code>.
*/
public String getParent() {
return parent;
}
/**
* Return a <code>Style</code> object corresponding to this one, but with
* all of the inherited information from parent <code>Style</code>
* objects filled in. The object returned will be a new object, not a
* reference to this object, even if it does not need any information
* added.
*
* @return A resolved <code>Style</code> object in which to look up
* ancestors.
*/
public Style getResolved() {
return new Style(name, family, parent, sc);
}
/**
* Write a <code>Node</code> in <code>parentDoc</code>
* representing this <code>Style</code>. Note that the
* <code>Node</code> is returned unconnected.
*
* @param parentDoc Document to which new <code>Node</code> will
* belong.
* @param name Name to use for new <code>Node</code>.
*/
public Node createNode(org.w3c.dom.Document parentDoc, String name) {
// DJP: write this! Should call writeAttributes()
return null;
}
/**
* Write this <code>Style</code> object's attributes to the given
* <code>Node</code>. This may involve writing child
* <code>Node</code> objects as well. This is similar to the
* <code>writeNode</code> method, but the <code>Node</code>
* already exists, and this does <b>not</b> write the name,
* family, and parent attributes, which are assumed to already
* exist in the <code>Node</code>.
*
* @param node The <code>Node</code> to add style attributes.
*/
public void writeAttributes(Node node) {
}
/**
* Return true if <code>Style</code> is a subset of this one. Note
* that this will return true even if <code>Style</code> is less
* specific than this <code>Style</code>, so long as it does not
* contradict this <code>Style</code> in any way.
*
* This always returns true since only subclasses of
* <code>Style</code> contain any actual <code>Style</code>
* information.
*
* @param style The <code>Style</code> to check
*
* @return true if the <code>Style</code> is a subset, false otherwise.
*/
public boolean isSubset(Style style) {
return true;
}
}
| 3,019 |
435 | <reponame>amaajemyfren/data<gh_stars>100-1000
{
"alias": "video/1480/python-web-sprints",
"category": "PyCon DE 2012",
"copyright_text": "",
"description": "",
"duration": null,
"id": 1480,
"language": "deu",
"quality_notes": "",
"recorded": "2012-11-01",
"slug": "python-web-sprints",
"speakers": [
"<NAME>",
"<NAME>"
],
"summary": "Seit April 2011 finden regelm\u00e4\u00dfig dreit\u00e4gige Sprints zu verschiedenen\nThemen rund um Python-Web Anwendungen statt. Dabei geht es meist nicht\num die schnelle Implementierung spezifischer Features sondern um\ngrundlegende, von den Beteiligten mitgetragene Konzepte. Meist wird auch\nnicht an einzelnen Komponenten, sondern an einer Toolchain zur L\u00f6sung\nkonkreter Probleme gearbeitet. Schlie\u00dflich werden die Ergebnisse in\neiner Anwenderdokumentation festgehalten.\n\nDie Initiatoren des Sprints, <NAME> und <NAME>, stellen\ndas Konzept der Sprint-Reihe vor und gehen auch die Ergebnisse der\nbisherigen Sprints ein. Die Themen der bisherigen Sprints waren:\nPyPI-Mirroring und Performance-Optimierungen, Sicherheit und Datenschutz\nbei Web-Anwendungen sowie Betrieb und Deployment von\nPython-Webanwendungen.\n",
"tags": [
"sprint",
"web-anwendungen"
],
"thumbnail_url": "https://i4.ytimg.com/vi/sX5zuqomc14/hqdefault.jpg",
"title": "Python-Web-Sprints",
"videos": [
{
"type": "mp4",
"url": "http://s3.us.archive.org/nextdayvideo/pyconde/pyconde2012/Vortrag_PythonWebSprints.mp4?Signature=kmgV9ltUeZT%2B4SIor4Lqna770PQ%3D&Expires=1352081375&AWSAccessKeyId=<KEY>"
},
{
"length": 0,
"type": "youtube",
"url": "https://www.youtube.com/watch?v=sX5zuqomc14"
}
]
}
| 742 |
2,329 | package fields;
public class Ja extends Ia {
public int getField() {
return field;
}
public void setField(int newvalue) {
field = newvalue;
}
}
| 56 |
337 | <gh_stars>100-1000
//statement
Character i = 10; | 17 |
4,054 | // Copyright Yahoo. Licensed under the terms of the Apache 2.0 license. See LICENSE in the project root.
package com.yahoo.vespa.hosted.controller.api.integration.noderepository;
import com.fasterxml.jackson.annotation.JsonCreator;
import com.fasterxml.jackson.annotation.JsonIgnoreProperties;
import com.fasterxml.jackson.annotation.JsonInclude;
import com.fasterxml.jackson.annotation.JsonProperty;
/**
* Patchable data under Application
*
* @author bratseth
*/
@JsonIgnoreProperties(ignoreUnknown = true)
@JsonInclude(JsonInclude.Include.NON_NULL)
public class ApplicationPatch {
@JsonProperty
private final Double currentReadShare;
@JsonProperty
private final Double maxReadShare;
@JsonCreator
public ApplicationPatch(@JsonProperty("currentReadShare") Double currentReadShare,
@JsonProperty("maxReadShare") Double maxReadShare) {
this.currentReadShare = currentReadShare;
this.maxReadShare = maxReadShare;
}
public Double getCurrentReadShare() { return currentReadShare; }
public Double getMaxReadShare() { return maxReadShare; }
}
| 381 |
852 | import FWCore.ParameterSet.Config as cms
#
# Strip calib
#
from CalibTracker.Configuration.SiStripCabling.SiStripCabling_Frontier_cff import *
from CalibTracker.Configuration.SiStripGain.SiStripGain_Frontier_cff import *
from CalibTracker.Configuration.SiStripLorentzAngle.SiStripLorentzAngle_Frontier_cff import *
from CalibTracker.Configuration.SiStripNoise.SiStripNoise_Frontier_cff import *
from CalibTracker.Configuration.SiStripPedestals.SiStripPedestals_Fake_cff import *
from CalibTracker.Configuration.SiStripQuality.SiStripQuality_Fake_cff import *
| 189 |
421 |
// <Snippet3>
#using <system.dll>
using namespace System;
public ref class HelloServiceClass: public MarshalByRefObject
{
private:
static int n_instance;
public:
HelloServiceClass()
{
n_instance++;
Console::WriteLine( "{0} has been created. Instance # = {1}", this->GetType()->Name, n_instance );
}
~HelloServiceClass()
{
Console::WriteLine( "Destroyed instance {0} of HelloServiceClass.", n_instance );
n_instance--;
}
String^ HelloMethod( String^ name )
{
// Reports that the method was called.
Console::WriteLine();
Console::WriteLine( "Called HelloMethod on instance {0} with the '{1}' parameter.", n_instance, name );
// Calculates and returns the result to the client.
return String::Format( "Hi there {0}", name );
}
};
// </Snippet3>
| 342 |
1,455 | <filename>src/main/java/org/springframework/data/redis/support/atomic/RedisAtomicDouble.java
/*
* Copyright 2013-2021 the original author or authors.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* https://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.springframework.data.redis.support.atomic;
import java.io.Serializable;
import java.util.Date;
import java.util.concurrent.TimeUnit;
import java.util.function.DoubleBinaryOperator;
import java.util.function.DoubleUnaryOperator;
import org.springframework.dao.DataRetrievalFailureException;
import org.springframework.data.redis.connection.DataType;
import org.springframework.data.redis.connection.RedisConnectionFactory;
import org.springframework.data.redis.core.BoundKeyOperations;
import org.springframework.data.redis.core.RedisOperations;
import org.springframework.data.redis.core.RedisTemplate;
import org.springframework.data.redis.core.ValueOperations;
import org.springframework.data.redis.serializer.GenericToStringSerializer;
import org.springframework.data.redis.serializer.RedisSerializer;
import org.springframework.lang.Nullable;
import org.springframework.util.Assert;
/**
* Atomic double backed by Redis. Uses Redis atomic increment/decrement and watch/multi/exec operations for CAS
* operations.
*
* @author <NAME>
* @author <NAME>
* @author <NAME>
* @author <NAME>
* @author <NAME>
* @author <NAME>
*/
public class RedisAtomicDouble extends Number implements Serializable, BoundKeyOperations<String> {
private static final long serialVersionUID = 1L;
private volatile String key;
private final ValueOperations<String, Double> operations;
private final RedisOperations<String, Double> generalOps;
/**
* Constructs a new {@link RedisAtomicDouble} instance. Uses the value existing in Redis or {@code 0} if none is
* found.
*
* @param redisCounter Redis key of this counter.
* @param factory connection factory.
*/
public RedisAtomicDouble(String redisCounter, RedisConnectionFactory factory) {
this(redisCounter, factory, null);
}
/**
* Constructs a new {@link RedisAtomicDouble} instance with a {@code initialValue} that overwrites the existing value.
*
* @param redisCounter Redis key of this counter.
* @param factory connection factory.
* @param initialValue initial value to set.
*/
public RedisAtomicDouble(String redisCounter, RedisConnectionFactory factory, double initialValue) {
this(redisCounter, factory, Double.valueOf(initialValue));
}
private RedisAtomicDouble(String redisCounter, RedisConnectionFactory factory, @Nullable Double initialValue) {
Assert.hasText(redisCounter, "a valid counter name is required");
Assert.notNull(factory, "a valid factory is required");
RedisTemplate<String, Double> redisTemplate = new RedisTemplate<>();
redisTemplate.setKeySerializer(RedisSerializer.string());
redisTemplate.setValueSerializer(new GenericToStringSerializer<>(Double.class));
redisTemplate.setExposeConnection(true);
redisTemplate.setConnectionFactory(factory);
redisTemplate.afterPropertiesSet();
this.key = redisCounter;
this.generalOps = redisTemplate;
this.operations = generalOps.opsForValue();
if (initialValue == null) {
initializeIfAbsent();
} else {
set(initialValue);
}
}
/**
* Constructs a new {@link RedisAtomicDouble} instance. Uses the value existing in Redis or 0 if none is found.
*
* @param redisCounter Redis key of this counter.
* @param template the template.
* @see #RedisAtomicDouble(String, RedisConnectionFactory, double)
*/
public RedisAtomicDouble(String redisCounter, RedisOperations<String, Double> template) {
this(redisCounter, template, null);
}
/**
* Constructs a new {@link RedisAtomicDouble} instance with a {@code initialValue} that overwrites the existing value
* at {@code redisCounter}.
* <p>
* Note: You need to configure the given {@code template} with appropriate {@link RedisSerializer} for the key and
* value.
* <p>
* As an alternative one could use the {@link #RedisAtomicDouble(String, RedisConnectionFactory, Double)} constructor
* which uses appropriate default serializers.
*
* @param redisCounter Redis key of this counter.
* @param template the template
* @param initialValue initial value to set if the Redis key is absent.
*/
public RedisAtomicDouble(String redisCounter, RedisOperations<String, Double> template, double initialValue) {
this(redisCounter, template, Double.valueOf(initialValue));
}
private RedisAtomicDouble(String redisCounter, RedisOperations<String, Double> template,
@Nullable Double initialValue) {
Assert.hasText(redisCounter, "a valid counter name is required");
Assert.notNull(template, "a valid template is required");
Assert.notNull(template.getKeySerializer(), "a valid key serializer in template is required");
Assert.notNull(template.getValueSerializer(), "a valid value serializer in template is required");
this.key = redisCounter;
this.generalOps = template;
this.operations = generalOps.opsForValue();
if (initialValue == null) {
initializeIfAbsent();
} else {
set(initialValue);
}
}
private void initializeIfAbsent() {
operations.setIfAbsent(key, (double) 0);
}
/**
* Get the current value.
*
* @return the current value.
*/
public double get() {
Double value = operations.get(key);
if (value != null) {
return value;
}
throw new DataRetrievalFailureException(String.format("The key '%s' seems to no longer exist.", key));
}
/**
* Set to the given value.
*
* @param newValue the new value.
*/
public void set(double newValue) {
operations.set(key, newValue);
}
/**
* Set to the given value and return the old value.
*
* @param newValue the new value.
* @return the previous value.
*/
public double getAndSet(double newValue) {
Double value = operations.getAndSet(key, newValue);
return value != null ? value : 0;
}
/**
* Atomically set the value to the given updated value if the current value {@code ==} the expected value.
*
* @param expect the expected value.
* @param update the new value.
* @return {@literal true} if successful. {@literal false} indicates that the actual value was not equal to the
* expected value.
*/
public boolean compareAndSet(double expect, double update) {
return generalOps.execute(new CompareAndSet<>(this::get, this::set, key, expect, update));
}
/**
* Atomically increment by one the current value.
*
* @return the previous value.
*/
public double getAndIncrement() {
return incrementAndGet() - 1.0;
}
/**
* Atomically decrement by one the current value.
*
* @return the previous value.
*/
public double getAndDecrement() {
return decrementAndGet() + 1.0;
}
/**
* Atomically add the given value to current value.
*
* @param delta the value to add.
* @return the previous value.
*/
public double getAndAdd(double delta) {
return addAndGet(delta) - delta;
}
/**
* Atomically update the current value using the given {@link DoubleUnaryOperator update function}.
*
* @param updateFunction the function which calculates the value to set. Should be a pure function (no side effects),
* because it will be applied several times if update attempts fail due to concurrent calls. Must not be
* {@literal null}.
* @return the previous value.
* @since 2.2
*/
public double getAndUpdate(DoubleUnaryOperator updateFunction) {
Assert.notNull(updateFunction, "Update function must not be null!");
double previousValue, newValue;
do {
previousValue = get();
newValue = updateFunction.applyAsDouble(previousValue);
} while (!compareAndSet(previousValue, newValue));
return previousValue;
}
/**
* Atomically update the current value using the given {@link DoubleBinaryOperator accumulator function}. The new
* value is calculated by applying the accumulator function to the current value and the given {@code updateValue}.
*
* @param updateValue the value which will be passed into the accumulator function.
* @param accumulatorFunction the function which calculates the value to set. Should be a pure function (no side
* effects), because it will be applied several times if update attempts fail due to concurrent calls. Must
* not be {@literal null}.
* @return the previous value.
* @since 2.2
*/
public double getAndAccumulate(double updateValue, DoubleBinaryOperator accumulatorFunction) {
Assert.notNull(accumulatorFunction, "Accumulator function must not be null!");
double previousValue, newValue;
do {
previousValue = get();
newValue = accumulatorFunction.applyAsDouble(previousValue, updateValue);
} while (!compareAndSet(previousValue, newValue));
return previousValue;
}
/**
* Atomically increment by one the current value.
*
* @return the updated value.
*/
public double incrementAndGet() {
return operations.increment(key, 1.0);
}
/**
* Atomically decrement by one the current value.
*
* @return the updated value.
*/
public double decrementAndGet() {
return operations.increment(key, -1.0);
}
/**
* Atomically add the given value to current value.
*
* @param delta the value to add.
* @return the updated value.
*/
public double addAndGet(double delta) {
return operations.increment(key, delta);
}
/**
* Atomically update the current value using the given {@link DoubleUnaryOperator update function}.
*
* @param updateFunction the function which calculates the value to set. Should be a pure function (no side effects),
* because it will be applied several times if update attempts fail due to concurrent calls. Must not be
* {@literal null}.
* @return the updated value.
* @since 2.2
*/
public double updateAndGet(DoubleUnaryOperator updateFunction) {
Assert.notNull(updateFunction, "Update function must not be null!");
double previousValue, newValue;
do {
previousValue = get();
newValue = updateFunction.applyAsDouble(previousValue);
} while (!compareAndSet(previousValue, newValue));
return newValue;
}
/**
* Atomically update the current value using the given {@link DoubleBinaryOperator accumulator function}. The new
* value is calculated by applying the accumulator function to the current value and the given {@code updateValue}.
*
* @param updateValue the value which will be passed into the accumulator function.
* @param accumulatorFunction the function which calculates the value to set. Should be a pure function (no side
* effects), because it will be applied several times if update attempts fail due to concurrent calls. Must
* not be {@literal null}.
* @return the updated value.
* @since 2.2
*/
public double accumulateAndGet(double updateValue, DoubleBinaryOperator accumulatorFunction) {
Assert.notNull(accumulatorFunction, "Accumulator function must not be null!");
double previousValue, newValue;
do {
previousValue = get();
newValue = accumulatorFunction.applyAsDouble(previousValue, updateValue);
} while (!compareAndSet(previousValue, newValue));
return newValue;
}
/**
* @return the String representation of the current value.
*/
@Override
public String toString() {
return Double.toString(get());
}
/*
* (non-Javadoc)
* @see org.springframework.data.redis.core.BoundKeyOperations#getKey()
*/
@Override
public String getKey() {
return key;
}
/*
* (non-Javadoc)
* @see org.springframework.data.redis.core.BoundKeyOperations#getType()
*/
@Override
public DataType getType() {
return DataType.STRING;
}
/*
* (non-Javadoc)
* @see org.springframework.data.redis.core.BoundKeyOperations#getExpire()
*/
@Override
public Long getExpire() {
return generalOps.getExpire(key);
}
/*
* (non-Javadoc)
* @see org.springframework.data.redis.core.BoundKeyOperations#expire(long, java.util.concurrent.TimeUnit)
*/
@Override
public Boolean expire(long timeout, TimeUnit unit) {
return generalOps.expire(key, timeout, unit);
}
/*
* (non-Javadoc)
* @see org.springframework.data.redis.core.BoundKeyOperations#expireAt(java.util.Date)
*/
@Override
public Boolean expireAt(Date date) {
return generalOps.expireAt(key, date);
}
/*
* (non-Javadoc)
* @see org.springframework.data.redis.core.BoundKeyOperations#persist()
*/
@Override
public Boolean persist() {
return generalOps.persist(key);
}
/*
* (non-Javadoc)
* @see org.springframework.data.redis.core.BoundKeyOperations#rename(java.lang.Object)
*/
@Override
public void rename(String newKey) {
generalOps.rename(key, newKey);
key = newKey;
}
/*
* (non-Javadoc)
* @see java.lang.Number#intValue()
*/
@Override
public int intValue() {
return (int) get();
}
/*
* (non-Javadoc)
* @see java.lang.Number#longValue()
*/
@Override
public long longValue() {
return (long) get();
}
/*
* (non-Javadoc)
* @see java.lang.Number#floatValue()
*/
@Override
public float floatValue() {
return (float) get();
}
/*
* (non-Javadoc)
* @see java.lang.Number#doubleValue()
*/
@Override
public double doubleValue() {
return get();
}
}
| 4,310 |
930 | package com.foxinmy.weixin4j.mp.component;
import java.io.Serializable;
import java.util.Date;
import javax.xml.bind.annotation.XmlAccessType;
import javax.xml.bind.annotation.XmlAccessorType;
import javax.xml.bind.annotation.XmlElement;
import javax.xml.bind.annotation.XmlRootElement;
import javax.xml.bind.annotation.XmlTransient;
/**
* 组件消息
*
* @className ComponentMessage
* @author jinyu(<EMAIL>)
* @date 2016年7月5日
* @since JDK 1.6
*/
@XmlRootElement(name = "xml")
@XmlAccessorType(XmlAccessType.FIELD)
public class ComponentMessage implements Serializable {
private static final long serialVersionUID = -7243616276403632118L;
/**
* 第三方平台appid
*/
@XmlElement(name = "AppId")
private String appId;
/**
* 事件类型
*/
@XmlElement(name = "InfoType")
private String eventType;
/**
* 时间戳
*/
@XmlElement(name = "CreateTime")
private long createTime;
/**
* Ticket内容
*/
@XmlElement(name = "ComponentVerifyTicket")
private String verifyTicket;
/**
* 授权方的Appid
*/
@XmlElement(name = "AuthorizerAppid")
private String authAppId;
/**
* 授权码,可用于换取公众号的接口调用凭据
*/
@XmlElement(name = "AuthorizationCode")
private String authCode;
/**
* 授权码过期时间
*/
@XmlElement(name = "AuthorizationCodeExpiredTime")
private long authCodeExpiredTime;
public String getAppId() {
return appId;
}
public String getEventType() {
return eventType;
}
@XmlTransient
public ComponentEventType getFormatEventType() {
return ComponentEventType.valueOf(eventType);
}
public long getCreateTime() {
return createTime;
}
@XmlTransient
public Date getFormatCreateTime() {
return createTime > 0l ? new Date(createTime * 1000l) : null;
}
public String getVerifyTicket() {
return verifyTicket;
}
public String getAuthAppId() {
return authAppId;
}
public String getAuthCode() {
return authCode;
}
public long getAuthCodeExpiredTime() {
return authCodeExpiredTime;
}
@XmlTransient
public Date getFormatAuthCodeExpiredTime() {
return authCodeExpiredTime > 0l ? new Date(authCodeExpiredTime * 1000l) : null;
}
@Override
public String toString() {
return "ComponentMessage [appId=" + appId + ", eventType=" + eventType + ", createTime=" + createTime
+ ", verifyTicket=" + verifyTicket + ", authAppId=" + authAppId + ", authCode=" + authCode
+ ", authCodeExpiredTime=" + authCodeExpiredTime + "]";
}
}
| 943 |
7,137 | package io.onedev.server.util.concurrent;
import java.util.concurrent.FutureTask;
public class PrioritizedFutureTask<T> extends FutureTask<T>
implements PriorityAware, Comparable<PriorityAware> {
private final int priority;
public PrioritizedFutureTask(PrioritizedCallable<T> callable) {
super(callable);
this.priority = callable.getPriority();
}
public PrioritizedFutureTask(PrioritizedRunnable runnable, T value) {
super(runnable, value);
this.priority = runnable.getPriority();
}
@Override
public int getPriority() {
return priority;
}
@Override
public int compareTo(PriorityAware o) {
return priority - o.getPriority();
}
}
| 229 |
1,626 | <filename>test/sample.cpp
//
// Copyright (c) 2016, Scientific Toolworks, Inc.
//
// This software is licensed under the MIT License. The LICENSE.md file
// describes the conditions under which this software may be distributed.
//
// Author: <NAME>
//
#include "Test.h"
#include <QMainWindow>
using namespace Test;
using namespace QTest;
class TestSample : public QObject
{
Q_OBJECT
private slots:
void initTestCase();
void sample();
void cleanupTestCase();
private:
int inputDelay = 99;
int closeDelay = 9999;
ScratchRepository *mRepo = nullptr;
QMainWindow *mWindow = nullptr;
};
void TestSample::initTestCase()
{
mRepo = new ScratchRepository;
mWindow = new QMainWindow;
//UiComponent *ui = new UiComponent(mRepo->repo(), mWindow);
//ui->show();
//QVERIFY(qWaitForWindowActive(ui));
}
void TestSample::sample()
{
//keyClick(ui, 'Qt::Key_Return', Qt::NoModifier, inputDelay);
}
void TestSample::cleanupTestCase()
{
qWait(closeDelay);
mWindow->close();
delete mRepo;
}
TEST_MAIN(TestSample)
#include "sample.moc"
| 381 |
14,668 | // Copyright 2020 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#ifndef CHROME_BROWSER_CHROMEOS_FILEAPI_FILE_CHANGE_SERVICE_OBSERVER_H_
#define CHROME_BROWSER_CHROMEOS_FILEAPI_FILE_CHANGE_SERVICE_OBSERVER_H_
#include "base/observer_list_types.h"
namespace storage {
class FileSystemURL;
} // namespace storage
namespace chromeos {
// An interface for an observer which receives `FileChangeService` events.
class FileChangeServiceObserver : public base::CheckedObserver {
public:
// Invoked when a file identified by `url` has been modified. Note that this
// will not get called on file creation or deletion.
virtual void OnFileModified(const storage::FileSystemURL& url) {}
// Invoked when a file has been copied from `src` to `dst`.
virtual void OnFileCopied(const storage::FileSystemURL& src,
const storage::FileSystemURL& dst) {}
// Invoked when a file has been moved from `src` to `dst`.
virtual void OnFileMoved(const storage::FileSystemURL& src,
const storage::FileSystemURL& dst) {}
};
} // namespace chromeos
#endif // CHROME_BROWSER_CHROMEOS_FILEAPI_FILE_CHANGE_SERVICE_OBSERVER_H_
| 433 |
2,342 | /***********************************************************************
Vczh Library++ 3.0
Developer: <NAME>(vczh)
GacUI::Control Styles::Windows7 Styles
Clases:
***********************************************************************/
#ifndef VCZH_PRESENTATION_CONTROLS_WIN7STYLES_GUIWIN7STYLESCOMMON
#define VCZH_PRESENTATION_CONTROLS_WIN7STYLES_GUIWIN7STYLESCOMMON
#include "../GuiCommonStyles.h"
namespace vl
{
namespace presentation
{
namespace win7
{
/***********************************************************************
Button Configuration
***********************************************************************/
struct Win7ButtonColors
{
Color borderColor;
Color backgroundColor;
Color g1;
Color g2;
Color g3;
Color g4;
Color textColor;
Color bulletLight;
Color bulletDark;
bool operator==(const Win7ButtonColors& colors)
{
return
borderColor == colors.borderColor &&
backgroundColor == colors.backgroundColor &&
g1 == colors.g1 &&
g2 == colors.g2 &&
g3 == colors.g3 &&
g4 == colors.g4 &&
textColor == colors.textColor &&
bulletLight == colors.bulletLight &&
bulletDark == colors.bulletDark;
}
bool operator!=(const Win7ButtonColors& colors)
{
return !(*this==colors);
}
void SetAlphaWithoutText(unsigned char a);
static Win7ButtonColors Blend(const Win7ButtonColors& c1, const Win7ButtonColors& c2, vint ratio, vint total);
static Win7ButtonColors ButtonNormal();
static Win7ButtonColors ButtonActive();
static Win7ButtonColors ButtonPressed();
static Win7ButtonColors ButtonDisabled();
static Win7ButtonColors ItemNormal();
static Win7ButtonColors ItemActive();
static Win7ButtonColors ItemSelected();
static Win7ButtonColors ItemDisabled();
static Win7ButtonColors CheckedNormal(bool selected);
static Win7ButtonColors CheckedActive(bool selected);
static Win7ButtonColors CheckedPressed(bool selected);
static Win7ButtonColors CheckedDisabled(bool selected);
static Win7ButtonColors ToolstripButtonNormal();
static Win7ButtonColors ToolstripButtonActive();
static Win7ButtonColors ToolstripButtonPressed();
static Win7ButtonColors ToolstripButtonSelected();
static Win7ButtonColors ToolstripButtonDisabled();
static Win7ButtonColors MenuBarButtonNormal();
static Win7ButtonColors MenuBarButtonActive();
static Win7ButtonColors MenuBarButtonPressed();
static Win7ButtonColors MenuBarButtonDisabled();
static Win7ButtonColors MenuItemButtonNormal();
static Win7ButtonColors MenuItemButtonNormalActive();
static Win7ButtonColors MenuItemButtonSelected();
static Win7ButtonColors MenuItemButtonSelectedActive();
static Win7ButtonColors MenuItemButtonDisabled();
static Win7ButtonColors MenuItemButtonDisabledActive();
static Win7ButtonColors TabPageHeaderNormal();
static Win7ButtonColors TabPageHeaderActive();
static Win7ButtonColors TabPageHeaderSelected();
};
struct Win7ButtonElements
{
elements::GuiSolidBorderElement* rectBorderElement;
elements::GuiRoundBorderElement* roundBorderElement;
elements::GuiSolidBackgroundElement* backgroundElement;
elements::GuiGradientBackgroundElement* topGradientElement;
elements::GuiGradientBackgroundElement* bottomGradientElement;
elements::GuiSolidLabelElement* textElement;
compositions::GuiBoundsComposition* textComposition;
compositions::GuiBoundsComposition* mainComposition;
compositions::GuiBoundsComposition* backgroundComposition;
compositions::GuiTableComposition* gradientComposition;
static Win7ButtonElements Create(bool verticalGradient, bool roundBorder, Alignment horizontal=Alignment::Center, Alignment vertical=Alignment::Center);
void Apply(const Win7ButtonColors& colors);
};
struct Win7CheckedButtonElements
{
elements::GuiSolidBorderElement* borderElement;
elements::GuiSolidBackgroundElement* backgroundElement;
elements::GuiGradientBackgroundElement* outerGradientElement;
elements::GuiGradientBackgroundElement* innerGradientElement;
elements::GuiSolidLabelElement* textElement;
elements::GuiSolidLabelElement* bulletCheckElement;
elements::GuiSolidBackgroundElement* bulletRadioElement;
compositions::GuiBoundsComposition* textComposition;
compositions::GuiBoundsComposition* mainComposition;
static Win7CheckedButtonElements Create(elements::ElementShape shape, bool backgroundVisible);
void Apply(const Win7ButtonColors& colors);
};
struct Win7MenuItemButtonElements
{
elements::GuiRoundBorderElement* borderElement;
elements::GuiSolidBackgroundElement* backgroundElement;
elements::GuiGradientBackgroundElement* gradientElement;
elements::Gui3DSplitterElement* splitterElement;
compositions::GuiCellComposition* splitterComposition;
elements::GuiImageFrameElement* imageElement;
elements::GuiSolidLabelElement* textElement;
compositions::GuiSharedSizeItemComposition* textComposition;
elements::GuiSolidLabelElement* shortcutElement;
compositions::GuiSharedSizeItemComposition* shortcutComposition;
elements::GuiPolygonElement* subMenuArrowElement;
compositions::GuiGraphicsComposition* subMenuArrowComposition;
compositions::GuiBoundsComposition* mainComposition;
static Win7MenuItemButtonElements Create();
void Apply(const Win7ButtonColors& colors);
void SetActive(bool value);
void SetSubMenuExisting(bool value);
};
struct Win7TextBoxColors
{
Color borderColor;
Color backgroundColor;
bool operator==(const Win7TextBoxColors& colors)
{
return
borderColor == colors.borderColor &&
backgroundColor == colors.backgroundColor;
}
bool operator!=(const Win7TextBoxColors& colors)
{
return !(*this==colors);
}
static Win7TextBoxColors Blend(const Win7TextBoxColors& c1, const Win7TextBoxColors& c2, vint ratio, vint total);
static Win7TextBoxColors Normal();
static Win7TextBoxColors Active();
static Win7TextBoxColors Focused();
static Win7TextBoxColors Disabled();
};
/***********************************************************************
Helper Functions
***********************************************************************/
extern Color Win7GetSystemWindowColor();
extern Color Win7GetSystemTabContentColor();
extern Color Win7GetSystemBorderColor();
extern Color Win7GetSystemBorderSinkColor();
extern Color Win7GetSystemBorderRaiseColor();
extern Color Win7GetSystemTextColor(bool enabled);
extern void Win7SetFont(elements::GuiSolidLabelElement* element, compositions::GuiBoundsComposition* composition, const FontProperties& fontProperties);
extern void Win7CreateSolidLabelElement(elements::GuiSolidLabelElement*& element, compositions::GuiBoundsComposition*& composition, Alignment horizontal, Alignment vertical);
extern void Win7CreateSolidLabelElement(elements::GuiSolidLabelElement*& element, compositions::GuiSharedSizeItemComposition*& composition, const WString& group, Alignment horizontal, Alignment vertical);
extern elements::text::ColorEntry Win7GetTextBoxTextColor();
}
}
}
#endif | 2,994 |
409 | <reponame>jtavara23/cassandra-workshop-series<filename>week4-AppDev-api/samples-codes/sample-kafka/kafka-dse-webui/src/main/java/com/datastax/demo/springdata/TickSpringDataRepository.java
package com.datastax.demo.springdata;
import com.datastax.demo.conf.DseConstants;
import com.datastax.demo.springdata.dto.TickData;
import com.datastax.demo.springdata.dto.TickDataPrimaryKey;
import org.springframework.data.cassandra.repository.Query;
import org.springframework.data.repository.reactive.ReactiveCrudRepository;
import reactor.core.publisher.Flux;
/** Retrieve all Tick from Table. */
public interface TickSpringDataRepository
extends ReactiveCrudRepository<TickData, TickDataPrimaryKey> {
@Query("SELECT * FROM " + DseConstants.STOCKS_TICKS + " WHERE symbol = ?0 LIMIT 100")
Flux<TickData> findBySymbol(String symbol);
@Query(
"SELECT * FROM "
+ DseConstants.STOCKS_TICKS
+ " WHERE symbol in ('BAC', 'DVMT','DIS','IBM','WMT') LIMIT 500")
Flux<TickData> findAllLastSymbols();
}
| 392 |
892 | <filename>advisories/unreviewed/2022/05/GHSA-3j75-qm7x-j2gc/GHSA-3j75-qm7x-j2gc.json
{
"schema_version": "1.2.0",
"id": "GHSA-3j75-qm7x-j2gc",
"modified": "2022-05-01T23:47:37Z",
"published": "2022-05-01T23:47:37Z",
"aliases": [
"CVE-2008-2200"
],
"details": "Multiple cross-site scripting (XSS) vulnerabilities in Maian Weblog 4.0 allow remote attackers to inject arbitrary web script or HTML via the (1) keywords parameter to admin/index.php in a blogs search action, the (2) msg_charset and (3) msg_header9 parameters to admin/inc/header.php, and the (4) keywords parameter to index.php in a search action.",
"severity": [
],
"affected": [
],
"references": [
{
"type": "ADVISORY",
"url": "https://nvd.nist.gov/vuln/detail/CVE-2008-2200"
},
{
"type": "WEB",
"url": "https://exchange.xforce.ibmcloud.com/vulnerabilities/42207"
},
{
"type": "WEB",
"url": "http://secunia.com/advisories/30060"
},
{
"type": "WEB",
"url": "http://securityreason.com/securityalert/3880"
},
{
"type": "WEB",
"url": "http://www.securityfocus.com/archive/1/491588/100/0/threaded"
},
{
"type": "WEB",
"url": "http://www.securityfocus.com/bid/29032"
}
],
"database_specific": {
"cwe_ids": [
"CWE-79"
],
"severity": "MODERATE",
"github_reviewed": false
}
} | 644 |
334 | from calvin.utilities.utils import enum
# Only TEMPORARY, TERMINATE and EXHAUST are used on highest level in actor and port manager
# TEMPORARY is used during migration
# TERMINATE is used for dereplication without preserving tokens
# EXHAUST is used for dereplication while preserving tokens
# The different exhaust disconnects are based on a higher level exhaust value
# EXHAUST and EXHAUST_PEER are used at actor port, endpoint and connection level
# EXHAUST_INPORT, EXHAUST_OUTPORT, EXHAUST_PEER_SEND, EXHAUST_PEER_RECV are used at queue level
DISCONNECT = enum('TEMPORARY', 'TERMINATE', 'EXHAUST', 'EXHAUST_PEER',
'EXHAUST_INPORT', 'EXHAUST_OUTPORT', 'EXHAUST_PEER_RECV', 'EXHAUST_PEER_SEND')
| 239 |
416 | /*
File: AVAudioChannelLayout.h
Framework: AVFoundation
Copyright (c) 2014-2015 Apple Inc. All Rights Reserved.
*/
#import <AVFAudio/AVAudioTypes.h>
#import <CoreAudio/CoreAudioTypes.h>
NS_ASSUME_NONNULL_BEGIN
/*!
@class AVAudioChannelLayout
@abstract A description of the roles of a set of audio channels.
@discussion
This object is a thin wrapper for the AudioChannelLayout structure, described
in <CoreAudio/CoreAudioTypes.h>.
*/
OS_EXPORT API_AVAILABLE(macos(10.10), ios(8.0), watchos(2.0), tvos(9.0))
@interface AVAudioChannelLayout : NSObject <NSSecureCoding> {
@private
AudioChannelLayoutTag _layoutTag;
AudioChannelLayout * _layout;
void *_reserved;
}
- (instancetype)init NS_UNAVAILABLE;
/*! @method initWithLayoutTag:
@abstract Initialize from a layout tag.
@param layoutTag
The tag.
@discussion
Returns nil if the tag is either kAudioChannelLayoutTag_UseChannelDescriptions or
kAudioChannelLayoutTag_UseChannelBitmap.
*/
- (nullable instancetype)initWithLayoutTag:(AudioChannelLayoutTag)layoutTag;
/*! @method initWithLayout:
@abstract Initialize from an AudioChannelLayout.
@param layout
The AudioChannelLayout.
@discussion
If the provided layout's tag is kAudioChannelLayoutTag_UseChannelDescriptions, this
initializer attempts to convert it to a more specific tag.
*/
- (instancetype)initWithLayout:(const AudioChannelLayout *)layout NS_DESIGNATED_INITIALIZER;
/*! @method isEqual:
@abstract Determine whether another AVAudioChannelLayout is exactly equal to this layout.
@param object
The AVAudioChannelLayout to compare against.
@discussion
The underlying AudioChannelLayoutTag and AudioChannelLayout are compared for equality.
*/
- (BOOL)isEqual:(id)object;
/*! @method layoutWithLayoutTag:
@abstract Create from a layout tag.
*/
+ (instancetype)layoutWithLayoutTag:(AudioChannelLayoutTag)layoutTag;
/*! @method layoutWithLayout:
@abstract Create from an AudioChannelLayout
*/
+ (instancetype)layoutWithLayout:(const AudioChannelLayout *)layout;
/*! @property layoutTag
@abstract The layout's tag. */
@property (nonatomic, readonly) AudioChannelLayoutTag layoutTag;
/*! @property layout
@abstract The underlying AudioChannelLayout. */
@property (nonatomic, readonly) const AudioChannelLayout *layout;
/*! @property channelCount
@abstract The number of channels of audio data.
*/
@property (nonatomic, readonly) AVAudioChannelCount channelCount;
@end
NS_ASSUME_NONNULL_END
| 791 |
496 | # Generated by Django 3.2 on 2021-08-17 23:00
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('posts', '0019_auto_20210528_1122'),
]
operations = [
migrations.RemoveField(
model_name='historicalpost',
name='label',
),
migrations.RemoveField(
model_name='post',
name='label',
),
]
| 204 |
634 | /****************************************************************
* Licensed to the Apache Software Foundation (ASF) under one *
* or more contributor license agreements. See the NOTICE file *
* distributed with this work for additional information *
* regarding copyright ownership. The ASF licenses this file *
* to you under the Apache License, Version 2.0 (the *
* "License"); you may not use this file except in compliance *
* with the License. You may obtain a copy of the License at *
* *
* http://www.apache.org/licenses/LICENSE-2.0 *
* *
* Unless required by applicable law or agreed to in writing, *
* software distributed under the License is distributed on an *
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY *
* KIND, either express or implied. See the License for the *
* specific language governing permissions and limitations *
* under the License. *
****************************************************************/
package org.apache.james.transport.mailets;
import java.util.regex.Pattern;
import com.google.common.base.MoreObjects;
import com.google.common.base.Objects;
public class ReplacingPattern {
private final Pattern matcher;
private final boolean repeat;
private final String substitution;
public ReplacingPattern(Pattern matcher, boolean repeat, String substitution) {
this.matcher = matcher;
this.repeat = repeat;
this.substitution = substitution;
}
public Pattern getMatcher() {
return matcher;
}
public boolean isRepeat() {
return repeat;
}
public String getSubstitution() {
return substitution;
}
@Override
public boolean equals(Object o) {
if (o instanceof ReplacingPattern) {
ReplacingPattern other = (ReplacingPattern) o;
return Objects.equal(matcher.pattern(), other.matcher.pattern())
&& Objects.equal(repeat, other.repeat)
&& Objects.equal(substitution, other.substitution);
}
return false;
}
@Override
public int hashCode() {
return Objects.hashCode(matcher, repeat, substitution);
}
@Override
public String toString() {
return MoreObjects.toStringHelper(this)
.add("matcher", matcher)
.add("repeat", repeat)
.add("substitution", substitution)
.toString();
}
} | 1,017 |
5,535 | /*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*
*/
#ifndef _PXFBRIDGE_H
#define _PXFBRIDGE_H
#include "libchurl.h"
#include "pxf_fragment.h"
#include "cdb/cdbvars.h"
#include "nodes/pg_list.h"
/*
* Execution state of a foreign scan using pxf_fdw.
*/
typedef struct PxfFdwScanState
{
CHURL_HEADERS churl_headers;
CHURL_HANDLE churl_handle;
StringInfoData uri;
ListCell *current_fragment;
Relation relation;
char *filter_str;
ExprState *quals;
List *fragments;
List *retrieved_attrs;
PxfOptions *options;
CopyState cstate;
} PxfFdwScanState;
/*
* Execution state of a foreign insert operation.
*/
typedef struct PxfFdwModifyState
{
CopyState cstate; /* state of writing to PXF */
CHURL_HANDLE churl_handle; /* curl handle */
CHURL_HEADERS churl_headers; /* curl headers */
StringInfoData uri; /* rest endpoint URI for modify */
Relation relation;
PxfOptions *options; /* FDW options */
} PxfFdwModifyState;
/* Clean up churl related data structures from the context */
void PxfBridgeCleanup(PxfFdwModifyState *context);
/* Sets up data before starting import */
void PxfBridgeImportStart(PxfFdwScanState *pxfsstate);
/* Sets up data before starting export */
void PxfBridgeExportStart(PxfFdwModifyState *pxfmstate);
/* Reads data from the PXF server into the given buffer of a given size */
int PxfBridgeRead(void *outbuf, int minlen, int maxlen, void *extra);
/* Writes data from the given buffer of a given size to the PXF server */
int PxfBridgeWrite(PxfFdwModifyState *context, char *databuf, int datalen);
#endif /* _PXFBRIDGE_H */
| 794 |
922 | # Copyright (c) 2021 - present / Neuralmagic, Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import os
from typing import List
import pytest
from sparseml.tensorflow_v1.optim import multi_step_lr_schedule, step_lr_schedule
from sparseml.tensorflow_v1.utils import tf_compat
@pytest.mark.skipif(
os.getenv("NM_ML_SKIP_TENSORFLOW_TESTS", False),
reason="Skipping tensorflow_v1 tests",
)
@pytest.mark.parametrize(
"start_step,end_step,init_lr,step_size,gamma",
[
(0, 100, 0.01, 1, 0.9),
(17, 100, 0.1, 15, 0.9),
(0, 200, 0.01, 15, 0.95),
(11, 200, 0.1, 7, 0.95),
(11, -1, 0.1, 3, 0.95),
],
)
def test_step_lr_schedule(
start_step: int, end_step: int, init_lr: float, step_size: int, gamma: float
):
with tf_compat.Graph().as_default():
global_step = tf_compat.placeholder(dtype=tf_compat.int64, shape=[])
learning_rate = step_lr_schedule(
global_step, start_step, end_step, step_size, init_lr, gamma
)
with tf_compat.Session() as sess:
expected = init_lr
for step in range(end_step + 10):
measured = sess.run(learning_rate, feed_dict={global_step: step})
if (
step - start_step
) % step_size == 0 and start_step < step <= end_step:
expected = expected * gamma
assert abs(measured - expected) < 1e-5
@pytest.mark.skipif(
os.getenv("NM_ML_SKIP_TENSORFLOW_TESTS", False),
reason="Skipping tensorflow_v1 tests",
)
@pytest.mark.parametrize(
"start_step,milestone_steps,init_lr,gamma",
[
(0, [5, 9, 13], 0.01, 0.9),
(13, [9, 13], 0.1, 0.9),
(0, [5, 9, 13], 0.01, 0.95),
(17, [9, 13], 0.1, 0.9),
],
)
def test_multi_step_lr_schedule(
start_step: int, milestone_steps: List[int], init_lr: float, gamma: float
):
with tf_compat.Graph().as_default():
global_step = tf_compat.placeholder(dtype=tf_compat.int64, shape=[])
learning_rate = multi_step_lr_schedule(
global_step, start_step, milestone_steps, init_lr, gamma
)
with tf_compat.Session() as sess:
for step in range(start_step + milestone_steps[-1] + 10):
measured = sess.run(learning_rate, feed_dict={global_step: step})
gammas = sum(
[1 for mile in milestone_steps if step >= mile + start_step]
)
expected = init_lr * gamma ** gammas
assert abs(measured - expected) < 1e-5
| 1,399 |
335 | <reponame>ford442/openmpt<filename>common/mptString.h
/*
* mptString.h
* ----------
* Purpose: Small string-related utilities, number and message formatting.
* Notes : Currently none.
* Authors: OpenMPT Devs
* The OpenMPT source code is released under the BSD license. Read LICENSE for more details.
*/
#pragma once
#include "openmpt/all/BuildSettings.hpp"
#include "mpt/base/alloc.hpp"
#include "mpt/base/span.hpp"
#include "mpt/string/types.hpp"
#include "mpt/string/utility.hpp"
#include "mptBaseTypes.h"
#include <algorithm>
#include <limits>
#include <string>
#include <string_view>
#include <cstring>
OPENMPT_NAMESPACE_BEGIN
namespace mpt
{
namespace String
{
template <typename Tstring, typename Tstring2, typename Tstring3>
inline Tstring Replace(Tstring str, const Tstring2 &oldStr, const Tstring3 &newStr)
{
return mpt::replace(str, oldStr, newStr);
}
} // namespace String
enum class Charset {
UTF8,
ASCII, // strictly 7-bit ASCII
ISO8859_1,
ISO8859_15,
CP850,
CP437,
CP437AMS,
CP437AMS2,
Windows1252,
#if defined(MPT_ENABLE_CHARSET_LOCALE)
Locale, // CP_ACP on windows, current C locale otherwise
#endif // MPT_ENABLE_CHARSET_LOCALE
};
// source code / preprocessor (i.e. # token)
inline constexpr Charset CharsetSource = Charset::ASCII;
// debug log files
inline constexpr Charset CharsetLogfile = Charset::UTF8;
// std::clog / std::cout / std::cerr
#if defined(MODPLUG_TRACKER) && MPT_OS_WINDOWS && defined(MPT_ENABLE_CHARSET_LOCALE)
inline constexpr Charset CharsetStdIO = Charset::Locale;
#else
inline constexpr Charset CharsetStdIO = Charset::UTF8;
#endif
// getenv
#if defined(MPT_ENABLE_CHARSET_LOCALE)
inline constexpr Charset CharsetEnvironment = Charset::Locale;
#else
inline constexpr Charset CharsetEnvironment = Charset::UTF8;
#endif
// std::exception::what()
#if defined(MPT_ENABLE_CHARSET_LOCALE)
inline constexpr Charset CharsetException = Charset::Locale;
#else
inline constexpr Charset CharsetException = Charset::UTF8;
#endif
// Checks if the std::string represents an UTF8 string.
// This is currently implemented as converting to std::wstring and back assuming UTF8 both ways,
// and comparing the result to the original string.
// Caveats:
// - can give false negatives because of possible unicode normalization during conversion
// - can give false positives if the 8bit encoding contains high-ascii only in valid utf8 groups
// - slow because of double conversion
bool IsUTF8(const std::string &str);
#if MPT_WSTRING_CONVERT
// Convert to a wide character string.
// The wide encoding is UTF-16 or UTF-32, based on sizeof(wchar_t).
// If str does not contain any invalid characters, this conversion is lossless.
// Invalid source bytes will be replaced by some replacement character or string.
inline std::wstring ToWide(const std::wstring &str) { return str; }
inline std::wstring ToWide(const wchar_t * str) { return (str ? std::wstring(str) : std::wstring()); }
std::wstring ToWide(Charset from, const std::string &str);
inline std::wstring ToWide(Charset from, const char * str) { return ToWide(from, str ? std::string(str) : std::string()); }
#if defined(MPT_ENABLE_CHARSET_LOCALE)
std::wstring ToWide(const mpt::lstring &str);
#endif // MPT_ENABLE_CHARSET_LOCALE
#endif
// Convert to a string encoded in the 'to'-specified character set.
// If str does not contain any invalid characters,
// this conversion will be lossless iff, and only iff,
// 'to' is UTF8.
// Invalid source bytes or characters that are not representable in the
// destination charset will be replaced by some replacement character or string.
#if MPT_WSTRING_CONVERT
std::string ToCharset(Charset to, const std::wstring &str);
inline std::string ToCharset(Charset to, const wchar_t * str) { return ToCharset(to, str ? std::wstring(str) : std::wstring()); }
#endif
std::string ToCharset(Charset to, Charset from, const std::string &str);
inline std::string ToCharset(Charset to, Charset from, const char * str) { return ToCharset(to, from, str ? std::string(str) : std::string()); }
#if defined(MPT_ENABLE_CHARSET_LOCALE)
std::string ToCharset(Charset to, const mpt::lstring &str);
#endif // MPT_ENABLE_CHARSET_LOCALE
#if defined(MPT_ENABLE_CHARSET_LOCALE)
#if MPT_WSTRING_CONVERT
mpt::lstring ToLocale(const std::wstring &str);
inline mpt::lstring ToLocale(const wchar_t * str) { return ToLocale(str ? std::wstring(str): std::wstring()); }
#endif
mpt::lstring ToLocale(Charset from, const std::string &str);
inline mpt::lstring ToLocale(Charset from, const char * str) { return ToLocale(from, str ? std::string(str): std::string()); }
inline mpt::lstring ToLocale(const mpt::lstring &str) { return str; }
#endif // MPT_ENABLE_CHARSET_LOCALE
#if MPT_OS_WINDOWS
#if MPT_WSTRING_CONVERT
mpt::winstring ToWin(const std::wstring &str);
inline mpt::winstring ToWin(const wchar_t * str) { return ToWin(str ? std::wstring(str): std::wstring()); }
#endif
mpt::winstring ToWin(Charset from, const std::string &str);
inline mpt::winstring ToWin(Charset from, const char * str) { return ToWin(from, str ? std::string(str): std::string()); }
#if defined(MPT_ENABLE_CHARSET_LOCALE)
mpt::winstring ToWin(const mpt::lstring &str);
#endif // MPT_ENABLE_CHARSET_LOCALE
#endif // MPT_OS_WINDOWS
#if defined(MPT_WITH_MFC)
#if !(MPT_WSTRING_CONVERT)
#error "MFC depends on MPT_WSTRING_CONVERT)"
#endif
// Convert to a MFC CString. The CString encoding depends on UNICODE.
// This should also be used when converting to TCHAR strings.
// If UNICODE is defined, this is a completely lossless operation.
inline CString ToCString(const CString &str) { return str; }
CString ToCString(const std::wstring &str);
inline CString ToCString(const wchar_t * str) { return ToCString(str ? std::wstring(str) : std::wstring()); }
CString ToCString(Charset from, const std::string &str);
inline CString ToCString(Charset from, const char * str) { return ToCString(from, str ? std::string(str) : std::string()); }
#if defined(MPT_ENABLE_CHARSET_LOCALE)
CString ToCString(const mpt::lstring &str);
mpt::lstring ToLocale(const CString &str);
#endif // MPT_ENABLE_CHARSET_LOCALE
#if MPT_OS_WINDOWS
mpt::winstring ToWin(const CString &str);
#endif // MPT_OS_WINDOWS
// Convert from a MFC CString. The CString encoding depends on UNICODE.
// This should also be used when converting from TCHAR strings.
// If UNICODE is defined, this is a completely lossless operation.
std::wstring ToWide(const CString &str);
std::string ToCharset(Charset to, const CString &str);
#endif // MPT_WITH_MFC
#define UC_(x) MPT_UCHAR(x)
#define UL_(x) MPT_ULITERAL(x)
#define U_(x) MPT_USTRING(x)
#if MPT_USTRING_MODE_WIDE
#if !(MPT_WSTRING_CONVERT)
#error "MPT_USTRING_MODE_WIDE depends on MPT_WSTRING_CONVERT)"
#endif
inline mpt::ustring ToUnicode(const std::wstring &str) { return str; }
inline mpt::ustring ToUnicode(const wchar_t * str) { return (str ? std::wstring(str) : std::wstring()); }
inline mpt::ustring ToUnicode(Charset from, const std::string &str) { return ToWide(from, str); }
inline mpt::ustring ToUnicode(Charset from, const char * str) { return ToUnicode(from, str ? std::string(str) : std::string()); }
#if defined(MPT_ENABLE_CHARSET_LOCALE)
inline mpt::ustring ToUnicode(const mpt::lstring &str) { return ToWide(str); }
#endif // MPT_ENABLE_CHARSET_LOCALE
#if defined(MPT_WITH_MFC)
inline mpt::ustring ToUnicode(const CString &str) { return ToWide(str); }
#endif // MFC
#else // !MPT_USTRING_MODE_WIDE
inline mpt::ustring ToUnicode(const mpt::ustring &str) { return str; }
#if MPT_WSTRING_CONVERT
mpt::ustring ToUnicode(const std::wstring &str);
inline mpt::ustring ToUnicode(const wchar_t * str) { return ToUnicode(str ? std::wstring(str) : std::wstring()); }
#endif
mpt::ustring ToUnicode(Charset from, const std::string &str);
inline mpt::ustring ToUnicode(Charset from, const char * str) { return ToUnicode(from, str ? std::string(str) : std::string()); }
#if defined(MPT_ENABLE_CHARSET_LOCALE)
mpt::ustring ToUnicode(const mpt::lstring &str);
#endif // MPT_ENABLE_CHARSET_LOCALE
#if defined(MPT_WITH_MFC)
mpt::ustring ToUnicode(const CString &str);
#endif // MPT_WITH_MFC
#endif // MPT_USTRING_MODE_WIDE
#if MPT_USTRING_MODE_WIDE
#if !(MPT_WSTRING_CONVERT)
#error "MPT_USTRING_MODE_WIDE depends on MPT_WSTRING_CONVERT)"
#endif
// nothing, std::wstring overloads will catch all stuff
#else // !MPT_USTRING_MODE_WIDE
#if MPT_WSTRING_CONVERT
std::wstring ToWide(const mpt::ustring &str);
#endif
std::string ToCharset(Charset to, const mpt::ustring &str);
#if defined(MPT_ENABLE_CHARSET_LOCALE)
mpt::lstring ToLocale(const mpt::ustring &str);
#endif // MPT_ENABLE_CHARSET_LOCALE
#if MPT_OS_WINDOWS
mpt::winstring ToWin(const mpt::ustring &str);
#endif // MPT_OS_WINDOWS
#if defined(MPT_WITH_MFC)
CString ToCString(const mpt::ustring &str);
#endif // MPT_WITH_MFC
#endif // MPT_USTRING_MODE_WIDE
// The MPT_UTF8 allows specifying UTF8 char arrays.
// The resulting type is mpt::ustring and the construction might require runtime translation,
// i.e. it is NOT generally available at compile time.
// Use explicit UTF8 encoding,
// i.e. U+00FC (LATIN SMALL LETTER U WITH DIAERESIS) would be written as "\xC3\xBC".
#define MPT_UTF8(x) mpt::ToUnicode(mpt::Charset::UTF8, x)
mpt::ustring ToUnicode(uint16 codepage, mpt::Charset fallback, const std::string &str);
char ToLowerCaseAscii(char c);
char ToUpperCaseAscii(char c);
std::string ToLowerCaseAscii(std::string s);
std::string ToUpperCaseAscii(std::string s);
int CompareNoCaseAscii(const char *a, const char *b, std::size_t n);
int CompareNoCaseAscii(std::string_view a, std::string_view b);
int CompareNoCaseAscii(const std::string &a, const std::string &b);
#if defined(MODPLUG_TRACKER)
mpt::ustring ToLowerCase(const mpt::ustring &s);
mpt::ustring ToUpperCase(const mpt::ustring &s);
#endif // MODPLUG_TRACKER
} // namespace mpt
// The AnyString types are meant to be used as function argument types only,
// and only during the transition phase to all-unicode strings in the whole codebase.
// Using an AnyString type as function argument avoids the need to overload a function for all the
// different string types that we currently have.
// Warning: These types will silently do charset conversions. Only use them when this can be tolerated.
// BasicAnyString is convertable to mpt::ustring and constructable from any string at all.
template <mpt::Charset charset = mpt::Charset::UTF8, bool tryUTF8 = true>
class BasicAnyString : public mpt::ustring
{
private:
static mpt::ustring From8bit(const std::string &str)
{
if constexpr(charset == mpt::Charset::UTF8)
{
return mpt::ToUnicode(mpt::Charset::UTF8, str);
} else
{
// auto utf8 detection
if constexpr(tryUTF8)
{
if(mpt::IsUTF8(str))
{
return mpt::ToUnicode(mpt::Charset::UTF8, str);
} else
{
return mpt::ToUnicode(charset, str);
}
} else
{
return mpt::ToUnicode(charset, str);
}
}
}
public:
// 8 bit
BasicAnyString(const char *str) : mpt::ustring(From8bit(str ? str : std::string())) { }
BasicAnyString(const std::string str) : mpt::ustring(From8bit(str)) { }
// locale
#if defined(MPT_ENABLE_CHARSET_LOCALE)
BasicAnyString(const mpt::lstring str) : mpt::ustring(mpt::ToUnicode(str)) { }
#endif // MPT_ENABLE_CHARSET_LOCALE
// unicode
BasicAnyString(const mpt::ustring &str) : mpt::ustring(str) { }
BasicAnyString(mpt::ustring &&str) : mpt::ustring(std::move(str)) { }
#if MPT_USTRING_MODE_UTF8 && MPT_WSTRING_CONVERT
BasicAnyString(const std::wstring &str) : mpt::ustring(mpt::ToUnicode(str)) { }
#endif
#if MPT_WSTRING_CONVERT
BasicAnyString(const wchar_t *str) : mpt::ustring(str ? mpt::ToUnicode(str) : mpt::ustring()) { }
#endif
// mfc
#if defined(MPT_WITH_MFC)
BasicAnyString(const CString &str) : mpt::ustring(mpt::ToUnicode(str)) { }
#endif // MPT_WITH_MFC
// fallback for custom string types
template <typename Tstring> BasicAnyString(const Tstring &str) : mpt::ustring(mpt::ToUnicode(str)) { }
template <typename Tstring> BasicAnyString(Tstring &&str) : mpt::ustring(mpt::ToUnicode(std::forward<Tstring>(str))) { }
};
// AnyUnicodeString is convertable to mpt::ustring and constructable from any unicode string,
class AnyUnicodeString : public mpt::ustring
{
public:
// locale
#if defined(MPT_ENABLE_CHARSET_LOCALE)
AnyUnicodeString(const mpt::lstring &str) : mpt::ustring(mpt::ToUnicode(str)) { }
#endif // MPT_ENABLE_CHARSET_LOCALE
// unicode
AnyUnicodeString(const mpt::ustring &str) : mpt::ustring(str) { }
AnyUnicodeString(mpt::ustring &&str) : mpt::ustring(std::move(str)) { }
#if MPT_USTRING_MODE_UTF8 && MPT_WSTRING_CONVERT
AnyUnicodeString(const std::wstring &str) : mpt::ustring(mpt::ToUnicode(str)) { }
#endif
#if MPT_WSTRING_CONVERT
AnyUnicodeString(const wchar_t *str) : mpt::ustring(str ? mpt::ToUnicode(str) : mpt::ustring()) { }
#endif
// mfc
#if defined(MPT_WITH_MFC)
AnyUnicodeString(const CString &str) : mpt::ustring(mpt::ToUnicode(str)) { }
#endif // MPT_WITH_MFC
// fallback for custom string types
template <typename Tstring> AnyUnicodeString(const Tstring &str) : mpt::ustring(mpt::ToUnicode(str)) { }
template <typename Tstring> AnyUnicodeString(Tstring &&str) : mpt::ustring(mpt::ToUnicode(std::forward<Tstring>(str))) { }
};
// AnyString
// Try to do the smartest auto-magic we can do.
#if defined(MPT_ENABLE_CHARSET_LOCALE)
using AnyString = BasicAnyString<mpt::Charset::Locale, true>;
#elif MPT_OS_WINDOWS
using AnyString = BasicAnyString<mpt::Charset::Windows1252, true>;
#else
using AnyString = BasicAnyString<mpt::Charset::ISO8859_1, true>;
#endif
// AnyStringLocale
// char-based strings are assumed to be in locale encoding.
#if defined(MPT_ENABLE_CHARSET_LOCALE)
using AnyStringLocale = BasicAnyString<mpt::Charset::Locale, false>;
#else
using AnyStringLocale = BasicAnyString<mpt::Charset::UTF8, false>;
#endif
// AnyStringUTF8orLocale
// char-based strings are tried in UTF8 first, if this fails, locale is used.
#if defined(MPT_ENABLE_CHARSET_LOCALE)
using AnyStringUTF8orLocale = BasicAnyString<mpt::Charset::Locale, true>;
#else
using AnyStringUTF8orLocale = BasicAnyString<mpt::Charset::UTF8, false>;
#endif
// AnyStringUTF8
// char-based strings are assumed to be in UTF8.
using AnyStringUTF8 = BasicAnyString<mpt::Charset::UTF8, false>;
OPENMPT_NAMESPACE_END
| 5,344 |
389 | <reponame>tcmoore32/sheer-madness<filename>gosu-test-api/src/main/java/gw/test/BaseRemoteTestClass.java<gh_stars>100-1000
package gw.test;
/*
* Copyright 2014 Guidewire Software, Inc.
*/
public class BaseRemoteTestClass extends TestClass {
protected BaseRemoteTestClass(String s, boolean shouldInit) {
super(s, shouldInit);
}
}
| 119 |
570 | <reponame>AmbientAI/amazon-kinesis-client<gh_stars>100-1000
/*
* Copyright 2019 Amazon.com, Inc. or its affiliates.
* Licensed under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package software.amazon.kinesis.leases;
import java.util.ArrayList;
import java.util.Collections;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
/**
* Shard Prioritization that prioritizes parent shards first.
* It also limits number of shards that will be available for initialization based on their depth.
* It doesn't make a lot of sense to work on a shard that has too many unfinished parents.
*/
public class ParentsFirstShardPrioritization implements
ShardPrioritization {
private static final SortingNode PROCESSING_NODE = new SortingNode(null, Integer.MIN_VALUE);
private final int maxDepth;
/**
* Creates ParentFirst prioritization with filtering based on depth of the shard.
* Shards that have depth > maxDepth will be ignored and will not be returned by this prioritization.
*
* @param maxDepth any shard that is deeper than max depth, will be excluded from processing
*/
public ParentsFirstShardPrioritization(int maxDepth) {
/* Depth 0 means that shard is completed or cannot be found,
* it is impossible to process such shards.
*/
if (maxDepth <= 0) {
throw new IllegalArgumentException("Max depth cannot be negative or zero. Provided value: " + maxDepth);
}
this.maxDepth = maxDepth;
}
@Override
public List<ShardInfo> prioritize(List<ShardInfo> original) {
Map<String, ShardInfo> shards = new HashMap<>();
for (ShardInfo shardInfo : original) {
shards.put(shardInfo.shardId(),
shardInfo);
}
Map<String, SortingNode> processedNodes = new HashMap<>();
for (ShardInfo shardInfo : original) {
populateDepth(shardInfo.shardId(),
shards,
processedNodes);
}
List<ShardInfo> orderedInfos = new ArrayList<>(original.size());
List<SortingNode> orderedNodes = new ArrayList<>(processedNodes.values());
Collections.sort(orderedNodes);
for (SortingNode sortingTreeNode : orderedNodes) {
// don't process shards with depth > maxDepth
if (sortingTreeNode.getDepth() <= maxDepth) {
orderedInfos.add(sortingTreeNode.shardInfo);
}
}
return orderedInfos;
}
private int populateDepth(String shardId,
Map<String, ShardInfo> shards,
Map<String, SortingNode> processedNodes) {
SortingNode processed = processedNodes.get(shardId);
if (processed != null) {
if (processed == PROCESSING_NODE) {
throw new IllegalArgumentException("Circular dependency detected. Shard Id "
+ shardId + " is processed twice");
}
return processed.getDepth();
}
ShardInfo shardInfo = shards.get(shardId);
if (shardInfo == null) {
// parent doesn't exist in our list, so this shard is root-level node
return 0;
}
if (shardInfo.isCompleted()) {
// we treat completed shards as 0-level
return 0;
}
// storing processing node to make sure we track progress and avoid circular dependencies
processedNodes.put(shardId, PROCESSING_NODE);
int maxParentDepth = 0;
for (String parentId : shardInfo.parentShardIds()) {
maxParentDepth = Math.max(maxParentDepth,
populateDepth(parentId,
shards,
processedNodes));
}
int currentNodeLevel = maxParentDepth + 1;
SortingNode previousValue = processedNodes.put(shardId,
new SortingNode(shardInfo,
currentNodeLevel));
if (previousValue != PROCESSING_NODE) {
throw new IllegalStateException("Validation failed. Depth for shardId " + shardId + " was populated twice");
}
return currentNodeLevel;
}
/**
* Class to store depth of shards during prioritization.
*/
private static class SortingNode implements
Comparable<SortingNode> {
private final ShardInfo shardInfo;
private final int depth;
public SortingNode(ShardInfo shardInfo,
int depth) {
this.shardInfo = shardInfo;
this.depth = depth;
}
public int getDepth() {
return depth;
}
@Override
public int compareTo(SortingNode o) {
return Integer.compare(depth,
o.depth);
}
}
}
| 2,145 |
315 | <reponame>FredrikBlomgren/aff3ct<filename>include/Tools/Algo/Draw_generator/Event_generator/MKL/Event_generator_MKL.hpp
/*!
* \file
* \brief Class tools::Event_generator_MKL.
*/
#ifdef AFF3CT_CHANNEL_MKL
#ifndef EVENT_GENERATOR_MKL_HPP
#define EVENT_GENERATOR_MKL_HPP
#include "Tools/types.h"
#include "Tools/Algo/Draw_generator/Event_generator/Event_generator.hpp"
namespace aff3ct
{
namespace tools
{
template <typename R = float, typename E = typename tools::matching_types<R>::B>
class Event_generator_MKL : public Event_generator<R,E>
{
private:
void* stream_state; // VSLStreamStatePtr* type
bool is_stream_alloc;
public:
explicit Event_generator_MKL(const int seed = 0);
virtual ~Event_generator_MKL();
virtual Event_generator_MKL<R,E>* clone() const;
virtual void set_seed(const int seed);
virtual void generate(E *draw, const unsigned length, const R event_probability);
};
}
}
#endif //EVENT_GENERATOR_MKL_HPP
#endif // MKL
| 370 |
575 | // Copyright (c) 2011 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "printing/printed_document.h"
#include "base/check_op.h"
#include "base/strings/utf_string_conversions.h"
#include "printing/metafile_skia.h"
#include "printing/page_number.h"
#include "printing/printed_page_win.h"
#include "printing/printing_context_win.h"
#include "printing/units.h"
#include "skia/ext/skia_utils_win.h"
namespace {
void SimpleModifyWorldTransform(HDC context,
int offset_x,
int offset_y,
float shrink_factor) {
XFORM xform = {0};
xform.eDx = static_cast<float>(offset_x);
xform.eDy = static_cast<float>(offset_y);
xform.eM11 = xform.eM22 = 1.f / shrink_factor;
BOOL res = ModifyWorldTransform(context, &xform, MWT_LEFTMULTIPLY);
DCHECK_NE(res, 0);
}
} // namespace
namespace printing {
void PrintedDocument::RenderPrintedPage(
const PrintedPage& page,
printing::NativeDrawingContext context) const {
#ifndef NDEBUG
{
// Make sure the page is from our list.
base::AutoLock lock(lock_);
DCHECK(&page == mutable_.pages_.find(page.page_number() - 1)->second.get());
}
#endif
DCHECK(context);
const PageSetup& page_setup = immutable_.settings_->page_setup_device_units();
gfx::Rect content_area = GetCenteredPageContentRect(
page_setup.physical_size(), page.page_size(), page.page_content_rect());
// Save the state to make sure the context this function call does not modify
// the device context.
int saved_state = SaveDC(context);
DCHECK_NE(saved_state, 0);
skia::InitializeDC(context);
{
// Save the state (again) to apply the necessary world transformation.
int saved_state_inner = SaveDC(context);
DCHECK_NE(saved_state_inner, 0);
// Setup the matrix to translate and scale to the right place. Take in
// account the actual shrinking factor.
// Note that the printing output is relative to printable area of the page.
// That is 0,0 is offset by PHYSICALOFFSETX/Y from the page.
SimpleModifyWorldTransform(
context, content_area.x() - page_setup.printable_area().x(),
content_area.y() - page_setup.printable_area().y(),
page.shrink_factor());
::StartPage(context);
bool played_back = page.metafile()->SafePlayback(context);
DCHECK(played_back);
::EndPage(context);
BOOL res = RestoreDC(context, saved_state_inner);
DCHECK_NE(res, 0);
}
BOOL res = RestoreDC(context, saved_state);
DCHECK_NE(res, 0);
}
bool PrintedDocument::RenderPrintedDocument(PrintingContext* context) {
if (context->NewPage() != PrintingContext::OK)
return false;
std::wstring device_name =
base::UTF16ToWide(immutable_.settings_->device_name());
{
base::AutoLock lock(lock_);
const MetafilePlayer* metafile = GetMetafile();
static_cast<PrintingContextWin*>(context)->PrintDocument(
device_name, *(static_cast<const MetafileSkia*>(metafile)));
}
return context->PageDone() == PrintingContext::OK;
}
} // namespace printing
| 1,185 |
2,083 | /*
* Copyright (C) 2015 Karumi.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.karumi.rosie.repository.datasource.paginated;
import com.karumi.rosie.repository.PaginatedCollection;
import com.karumi.rosie.repository.datasource.Identifiable;
import java.util.Collection;
/**
* Data source interface meant to be used only to persist page of data.
*
* @param <K> The class of the key used by this data source.
* @param <V> The class of the values stored into this data source.
*/
public interface PaginatedWriteableDataSource<K, V extends Identifiable<K>> {
/**
* Adds or update a page of values into this data source.
*
* @param page Page to be stored
* @param values Collection of values to be stored
* @param hasMore True whether the persisted page has more elements
*/
PaginatedCollection<V> addOrUpdatePage(Page page, Collection<V> values, boolean hasMore)
throws Exception;
/**
* Deletes all the pages stored in this data source.
*/
void deleteAll() throws Exception;
}
| 438 |
386 | <reponame>beru/libonnx<filename>src/default/Mean.c<gh_stars>100-1000
#include <onnx.h>
static int Mean_init(struct onnx_node_t * n)
{
if((n->ninput >= 1) && (n->noutput == 1))
return 1;
return 0;
}
static int Mean_exit(struct onnx_node_t * n)
{
return 1;
}
static int Mean_reshape(struct onnx_node_t * n)
{
struct onnx_tensor_t * y = n->outputs[0];
int i;
if(!onnx_tensor_reshape_identity(y, n->inputs[0], n->inputs[0]->type))
return 0;
for(i = 1; i < n->ninput; i++)
{
if(!onnx_tensor_reshape_multi_broadcast(y, y, n->inputs[i], y->type))
return 0;
}
return 1;
}
static void Mean_bfloat16(struct onnx_node_t * n)
{
struct onnx_tensor_t * y = n->outputs[0];
struct onnx_tensor_t * x;
uint16_t * py = (uint16_t *)y->datas;
uint16_t * px;
float sum;
size_t i, j, l;
for(i = 0, l = y->ndata; i < l; i++)
{
for(j = 0, sum = 0; j < n->ninput; j++)
{
x = n->inputs[j];
px = onnx_tensor_broadcast_map_address(x, y, i);
sum += bfloat16_to_float32(*px);
}
py[i] = float32_to_bfloat16(sum / n->ninput);
}
}
static void Mean_float16(struct onnx_node_t * n)
{
struct onnx_tensor_t * y = n->outputs[0];
struct onnx_tensor_t * x;
uint16_t * py = (uint16_t *)y->datas;
uint16_t * px;
float sum;
size_t i, j, l;
for(i = 0, l = y->ndata; i < l; i++)
{
for(j = 0, sum = 0; j < n->ninput; j++)
{
x = n->inputs[j];
px = onnx_tensor_broadcast_map_address(x, y, i);
sum += float16_to_float32(*px);
}
py[i] = float32_to_float16(sum / n->ninput);
}
}
static void Mean_float32(struct onnx_node_t * n)
{
struct onnx_tensor_t * y = n->outputs[0];
struct onnx_tensor_t * x;
float * py = (float *)y->datas;
float * px;
float sum;
size_t i, j, l;
for(i = 0, l = y->ndata; i < l; i++)
{
for(j = 0, sum = 0; j < n->ninput; j++)
{
x = n->inputs[j];
px = onnx_tensor_broadcast_map_address(x, y, i);
sum += *px;
}
py[i] = sum / n->ninput;
}
}
static void Mean_float64(struct onnx_node_t * n)
{
struct onnx_tensor_t * y = n->outputs[0];
struct onnx_tensor_t * x;
double * py = (double *)y->datas;
double * px;
double sum;
size_t i, j, l;
for(i = 0, l = y->ndata; i < l; i++)
{
for(j = 0, sum = 0; j < n->ninput; j++)
{
x = n->inputs[j];
px = onnx_tensor_broadcast_map_address(x, y, i);
sum += *px;
}
py[i] = sum / n->ninput;
}
}
void resolver_default_op_Mean(struct onnx_node_t * n)
{
if(n->opset >= 13)
{
switch(n->inputs[0]->type)
{
case ONNX_TENSOR_TYPE_BFLOAT16:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_bfloat16;
break;
case ONNX_TENSOR_TYPE_FLOAT16:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float16;
break;
case ONNX_TENSOR_TYPE_FLOAT32:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float32;
break;
case ONNX_TENSOR_TYPE_FLOAT64:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float64;
break;
default:
break;
}
}
else if(n->opset >= 8)
{
switch(n->inputs[0]->type)
{
case ONNX_TENSOR_TYPE_FLOAT16:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float16;
break;
case ONNX_TENSOR_TYPE_FLOAT32:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float32;
break;
case ONNX_TENSOR_TYPE_FLOAT64:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float64;
break;
default:
break;
}
}
else if(n->opset >= 6)
{
switch(n->inputs[0]->type)
{
case ONNX_TENSOR_TYPE_FLOAT16:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float16;
break;
case ONNX_TENSOR_TYPE_FLOAT32:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float32;
break;
case ONNX_TENSOR_TYPE_FLOAT64:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float64;
break;
default:
break;
}
}
else if(n->opset >= 1)
{
switch(n->inputs[0]->type)
{
case ONNX_TENSOR_TYPE_FLOAT16:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float16;
break;
case ONNX_TENSOR_TYPE_FLOAT32:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float32;
break;
case ONNX_TENSOR_TYPE_FLOAT64:
n->init = Mean_init;
n->exit = Mean_exit;
n->reshape = Mean_reshape;
n->operator = Mean_float64;
break;
default:
break;
}
}
}
| 2,428 |
1,405 | package a.a.a.a.a.a;
import java.io.InputStream;
import java.io.OutputStream;
public final class j implements o {
/* renamed from: a reason: collision with root package name */
private Class f20a;
private String b;
private Object c;
public j(String str) {
this.b = str;
}
@Override // a.a.a.a.a.a.o
public final void a() {
if (!h.a("com.ibm.mqttdirect.modules.local.bindings.LocalListener")) {
throw h.a(32103);
}
try {
this.f20a = Class.forName("com.ibm.mqttdirect.modules.local.bindings.LocalListener");
this.c = this.f20a.getMethod("connect", String.class).invoke(null, this.b);
} catch (Exception e) {
}
if (this.c == null) {
throw h.a(32103);
}
}
@Override // a.a.a.a.a.a.o
public final InputStream b() {
try {
return (InputStream) this.f20a.getMethod("getClientInputStream", new Class[0]).invoke(this.c, new Object[0]);
} catch (Exception e) {
return null;
}
}
@Override // a.a.a.a.a.a.o
public final OutputStream c() {
try {
return (OutputStream) this.f20a.getMethod("getClientOutputStream", new Class[0]).invoke(this.c, new Object[0]);
} catch (Exception e) {
return null;
}
}
@Override // a.a.a.a.a.a.o
public final void d() {
if (this.c != null) {
try {
this.f20a.getMethod("close", new Class[0]).invoke(this.c, new Object[0]);
} catch (Exception e) {
}
}
}
}
| 793 |
764 | <reponame>641589523/token-profile<filename>erc20/0x15F0EEDF9Ce24fc4b6826E590A8292CE5524a1DA.json
{"symbol": "DENA","address": "0x15F0EEDF9Ce24fc4b6826E590A8292CE5524a1DA","overview":{"en": ""},"email": "","website": "https://denations.com/","state": "NORMAL","links": {"blog": "https://smatoos.medium.com/","twitter": "https://twitter.com/SMATOOS_now","telegram": "https://t.me/smatoos","github": "https://github.com/denation-smatoos/denations"}} | 184 |
2,045 | '''
This is the benchmark of 4 different implementations
of rectified linear activation in Theano.
Two types of computations are tested w.r.t.
each implementation: fprop and grad.
Results: in seconds, float32 (details in the code)
Implementation, CPU (fprop, bprop), GPU (fprop, bprop), (final score)
a) ScalarRectifier: (2.32, 2.40) (1.36, 2.67) (8.75)
b) T.max(.0, x): (5.19, 3.65) (1.38, 2.38) (12.60)
c) x*(x>0.): (2.85, 2.84) (1.31, 2.91) (9.91)
d) T.switch(x<0., 0., x): (2.32, 1.41) (1.41, 2.84) (8.39)
Conlusion:
In terms of efficiency, d) > a) > c) > b)
'''
from __future__ import print_function
__authors__ = "<NAME> and <NAME>"
import theano
import theano.tensor as T
import numpy
import time
floatX = 'float32'
def relu(x):
"""
relu implementation with T.maximum
Parameters
----------
x: tensor variable
"""
return T.maximum(0.0, x)
def relu_(x):
"""
Alternative relu implementation
Parameters
----------
x: tensor variable
"""
return x * (x > 0)
def relu__(x):
"""
Alternative relu implementation. The most efficient one.
Parameters
----------
x: tensor variable
"""
return T.switch(x < 0., 0., x)
def test_scalar_rectifier():
"""
verify different implementations of relu
"""
x = T.fmatrix('inputs')
y1 = relu(x)
y3 = relu_(x)
y4 = relu__(x)
f1 = theano.function(inputs=[x], outputs=y1, name='benchmark_1_forward')
f3 = theano.function(inputs=[x], outputs=y3, name='benchmark_3_forward')
f4 = theano.function(inputs=[x], outputs=y4, name='benchmark_4_forward')
g1 = theano.function(inputs=[x], outputs=T.grad(y1.sum(),x),
name='benchmark_1_grad')
g3 = theano.function(inputs=[x], outputs=T.grad(y3.sum(),x),
name='benchmark_3_grad')
g4 = theano.function(inputs=[x], outputs=T.grad(y4.sum(),x),
name='benchmark_4_grad')
for i in range(10):
value = numpy.random.uniform(-1,1,size=(100,500)).astype(floatX)
numpy.testing.assert_array_equal(f1(value), f3(value),
err_msg='arrays not equal' )
numpy.testing.assert_array_equal(f1(value), f4(value),
err_msg='arrays not equal' )
numpy.testing.assert_array_equal(g1(value), g3(value),
err_msg='grad:arrays not equal' )
numpy.testing.assert_array_equal(g1(value), g4(value),
err_msg='grad:arrays not equal' )
def benchmark_relu():
"""
Benchmark the speed of different relu implementations.
Both fprop and grad are tested.
"""
x = T.ftensor4('inputs')
ops = [
relu_(x).sum(), # old
relu(x).sum(), # alter, short for alternative
relu__(x).sum(), # alter 2
T.grad(relu_(x).sum(),x), # grad_old
T.grad(relu(x).sum(),x), # grad_alter
T.grad(relu__(x).sum(),x), # grad_alter2
]
names = ['fprop_old', 'fprop_alter', 'fprop_alter2',
'grad_old', 'grad_alter', 'grad_alter2']
value = numpy.random.uniform(size=(512,32,32,100)).astype(floatX)
times = []
for op, name in zip(ops, names):
f = theano.function(inputs=[x], outputs=op, name=name)
n_loops = 10
t0 = time.time()
for i in range(n_loops):
f(value)
t1 = time.time()
benchmark = t1-t0
times.append(benchmark)
print(name)
theano.printing.debugprint(f, print_type=True)
print(names)
print(times)
if __name__ == '__main__':
benchmark_relu()
#test_scalar_rectifier()
| 1,780 |
649 | package net.thucydides.core.reports.integration;
import net.thucydides.core.annotations.Feature;
import net.thucydides.core.annotations.Story;
import net.thucydides.core.issues.IssueTracking;
import net.thucydides.core.issues.SystemPropertiesIssueTracking;
import net.thucydides.core.model.TestOutcome;
import net.thucydides.core.model.TestStep;
import net.thucydides.core.reports.AcceptanceTestReporter;
import net.thucydides.core.reports.html.HtmlAcceptanceTestReporter;
import net.thucydides.core.screenshots.ScreenshotAndHtmlSource;
import net.thucydides.core.util.ExtendedTemporaryFolder;
import net.thucydides.core.util.FileSystemUtils;
import net.thucydides.core.util.MockEnvironmentVariables;
import org.apache.commons.io.FileUtils;
import org.junit.Before;
import org.junit.Rule;
import java.io.File;
import java.io.IOException;
public class AbstractReportGenerationTest {
@Rule
public ExtendedTemporaryFolder temporaryDirectory = new ExtendedTemporaryFolder();
protected AcceptanceTestReporter reporter;
protected File outputDirectory;
protected MockEnvironmentVariables environmentVariables;
public AbstractReportGenerationTest() {
}
@Before
public void setupTestReporter() throws IOException {
environmentVariables = new MockEnvironmentVariables();
IssueTracking issueTracking = new SystemPropertiesIssueTracking(environmentVariables);
reporter = new HtmlAcceptanceTestReporter(environmentVariables, issueTracking);
outputDirectory = temporaryDirectory.newFolder();
reporter.setOutputDirectory(outputDirectory);
}
protected void recordStepWithScreenshot(final TestOutcome testOutcome,
final String stepName,
final String screenshot) throws IOException {
String screenshotResource = "/screenshots/" + screenshot;
File sourceFile = FileSystemUtils.getResourceAsFile(screenshotResource);// new File(sourcePath.getPath());
FileUtils.copyFileToDirectory(sourceFile, outputDirectory);
TestStep step = TestStepFactory.successfulTestStepCalled(stepName);
step.addScreenshot(new ScreenshotAndHtmlSource(new File(outputDirectory, screenshot), null));
testOutcome.recordStep(step);
}
protected class AUserStory {
}
@Story(WhenGeneratingAnHtmlReport.AUserStory.class)
protected class SomeTestScenario {
public void a_simple_test_case() {
}
;
public void should_do_this() {
}
;
public void should_do_that() {
}
;
}
@Feature
protected class AFeature {
class AUserStoryInAFeature {
}
;
}
@Story(WhenGeneratingAnHtmlReport.AFeature.AUserStoryInAFeature.class)
protected class SomeTestScenarioInAFeature {
public void should_do_this() {
}
;
public void should_do_that() {
}
;
}
} | 1,129 |
357 | <reponame>nicesu/wiki<filename>Programming/Java/Collection/StackDemo.java<gh_stars>100-1000
import java.util.*;
public class StackDemo {
public static void main(String args[])
{
Stack s = new Stack();
s.push(new Integer(0));
s.push(new Student(202));
s.push(new Integer(1)); //push in
System.out.println( (Integer)s.peek()); //get but not pop
System.out.println((Integer)s.pop()); //pop out
Student a =(Student)s.pop();
a.prt();
System.out.println((Integer)s.pop());
}
}
class Student
{
int no;
Student(int i) {no = i;}
void prt(){System.out.println(no);}
} | 275 |
666 | from lib import BaseTest
class PublishShow1Test(BaseTest):
"""
publish show: existing snapshot
"""
fixtureDB = True
fixturePool = True
fixtureCmds = [
"aptly snapshot create snap1 from mirror gnuplot-maverick",
"aptly publish snapshot -keyring=${files}/aptly.pub -secret-keyring=${files}/aptly.sec snap1",
]
runCmd = "aptly publish show maverick"
class PublishShow2Test(BaseTest):
"""
publish show: under prefix
"""
fixtureDB = True
fixturePool = True
fixtureCmds = [
"aptly snapshot create snap1 from mirror gnuplot-maverick",
"aptly publish snapshot -keyring=${files}/aptly.pub -secret-keyring=${files}/aptly.sec snap1 ppa/smira",
]
runCmd = "aptly publish show maverick ppa/smira"
class PublishShow3Test(BaseTest):
"""
publish show json: existing snapshot
"""
fixtureDB = True
fixturePool = True
fixtureCmds = [
"aptly snapshot create snap1 from mirror gnuplot-maverick",
"aptly publish snapshot -keyring=${files}/aptly.pub -secret-keyring=${files}/aptly.sec snap1",
]
runCmd = "aptly publish show -json maverick"
class PublishShow4Test(BaseTest):
"""
publish show json: under prefix
"""
fixtureDB = True
fixturePool = True
fixtureCmds = [
"aptly snapshot create snap1 from mirror gnuplot-maverick",
"aptly publish snapshot -keyring=${files}/aptly.pub -secret-keyring=${files}/aptly.sec snap1 ppa/smira",
]
runCmd = "aptly publish show -json maverick ppa/smira"
| 608 |
331 | <filename>Easy Challenges/Challenge 0020 Easy/solutions/solution.java
package dailyprogrammerchallenges;
import static java.lang.Math.sqrt;
import java.util.Scanner;
/**
*
* @author coderMato
*/
public class PrimeNumbers {
public static void main(String[] args){
int prime,i,j;
System.out.println("2");
for(i=3; i <= 2000; i++){
for(j=2; j <= (int) sqrt(i) + 1; j++){
if(i % j != 0 && j == (int) sqrt(i) + 1){
System.out.println(i);
}
if (i%j == 0){
break;
}
}
}
}
}
| 344 |
13,057 | <filename>src/main/java/org/mockito/plugins/InlineMockMaker.java
/*
* Copyright (c) 2019 Mockito contributors
* This program is made available under the terms of the MIT License.
*/
package org.mockito.plugins;
import org.mockito.MockitoFramework;
/**
* Extension to {@link MockMaker} for mock makers that changes inline method implementations
* and need keep track of created mock objects.
* <p>
* Mockito's default inline mock maker keeps track of created mock objects via weak reference map.
* This poses a risk of memory leaks in certain scenarios
* (issue <a href="https://github.com/mockito/mockito/pull/1619">#1619</a>).
* There is no clean way to tackle those problems at the moment.
* Hence, {@code InlineMockMaker} interface exposes methods to explicitly clear mock references.
* Those methods are called by {@link MockitoFramework#clearInlineMocks()}.
* When the user encounters a leak, he can mitigate the problem with {@link MockitoFramework#clearInlineMocks()}.
* <p>
* {@code InlineMockMaker} is for expert users and framework integrators, when custom inline mock maker is in use.
* If you have a custom {@link MockMaker} that keeps track of mock objects,
* please have your mock maker implement {@code InlineMockMaker} interface.
* This way, it can participate in {@link MockitoFramework#clearInlineMocks()} API.
*
* @since 2.25.0
*/
public interface InlineMockMaker extends MockMaker {
/**
* Clean up internal state for specified {@code mock}. You may assume there won't be any interaction to the specific
* mock after this is called.
*
* @param mock the mock instance whose internal state is to be cleaned.
* @since 2.25.0
*/
void clearMock(Object mock);
/**
* Cleans up internal state for all existing mocks. You may assume there won't be any interaction to mocks created
* previously after this is called.
*
* @since 2.25.0
*/
void clearAllMocks();
}
| 582 |
1,958 | package com.freetymekiyan.algorithms.level.hard;
/**
* The demons had captured the princess (P) and imprisoned her in the
* bottom-right corner of a dungeon. The dungeon consists of M x N rooms laid
* out in a 2D grid. Our valiant knight (K) was initially positioned in the
* top-left room and must fight his way through the dungeon to rescue the
* princess.
* <p>
* The knight has an initial health point represented by a positive integer. If
* at any point his health point drops to 0 or below, he dies immediately.
* <p>
* Some of the rooms are guarded by demons, so the knight loses health
* (negative integers) upon entering these rooms; other rooms are either empty
* (0's) or contain magic orbs that increase the knight's health (positive
* integers).
* <p>
* In order to reach the princess as quickly as possible, the knight decides to
* move only rightward or downward in each step.
* <p>
* Write a function to determine the knight's minimum initial health so that he
* is able to rescue the princess.
* <p>
* For example, given the dungeon below, the initial health of the knight must
* be at least 7 if he follows the optimal path RIGHT-> RIGHT -> DOWN -> DOWN.
* <p>
* -2(K) -3 3
* -5 -10 1
* 10 30 -5(P)
* <p>
* Notes:
* The knight's health has no upper bound.
* Any room can contain threats or power-ups, even the first room the knight
* enters and the bottom-right room where the princess is imprisoned.
* <p>
* Tags:DP, Binary Search
*/
class DungeonGame {
/**
* Build a table for the minimum hp needed to get to the bottom right
* Build from bottom right to get minimum from i, j to the end
* Instead of build from to-left, because it's hard to get correct relation
*/
public int calculateMinimumHP(int[][] dungeon) {
if (dungeon == null || dungeon.length == 0 || dungeon[0].length == 0) return 0;
int m = dungeon.length - 1;
int n = dungeon[0].length - 1;
dungeon[m][n] = Math.max(1 - dungeon[m][n], 1);
for (int i = m - 1; i >= 0; i--) dungeon[i][n] = Math.max(dungeon[i + 1][n] - dungeon[i][n], 1);
for (int j = n - 1; j >= 0; j--) dungeon[m][j] = Math.max(dungeon[m][j + 1] - dungeon[m][j], 1);
for (int i = m - 1; i >= 0; i--) {
for (int j = n - 1; j >= 0; j--) {
dungeon[i][j] = Math.max(Math.min(dungeon[i + 1][j], dungeon[i][j + 1]) - dungeon[i][j], 1);
}
}
return dungeon[0][0];
}
/**
* Why “from the bottom right corner to left top”?
* It depends on the way you formulate the problem. If you define a value
* in DP table d[i][j] as 'the minimum hp required to REACH (i, j) from (0,
* 0)", then the final answer should be d[nrows-1][ncols-1], and you need
* to start filling from the top left;
* However, in the reference answer provided with the question, dp[i][j] is
* defined as 'the minimum hp required to REACH (nrows-1, ncols-1) from (i,
* j)'. Here dp[0][0] is the final answer so we must fill from (nrows-1,
* ncols-1). For many other problems such as 'Minimum Path Sum', both
* formulation would work.
* However, in this problem, the former formulation will lead us to
* trouble, because it is very hard, if not impossible, to get d[i][j]
* based on d[i-1][j] and d[i][j-1].
*/
} | 1,107 |
2,338 | // RUN: %libomp-compile-and-run
#include <stdio.h>
#include <math.h>
#include "omp_testsuite.h"
#include "omp_my_sleep.h"
int test_omp_task_final()
{
int tids[NUM_TASKS];
int includedtids[NUM_TASKS];
int i;
int error = 0;
#pragma omp parallel
{
#pragma omp single
{
for (i = 0; i < NUM_TASKS; i++) {
/* First we have to store the value of the loop index in a new variable
* which will be private for each task because otherwise it will be overwritten
* if the execution of the task takes longer than the time which is needed to
* enter the next step of the loop!
*/
int myi;
myi = i;
#pragma omp task final(i>=10)
{
tids[myi] = omp_get_thread_num();
/* we generate included tasks for final tasks */
if(myi >= 10) {
int included = myi;
#pragma omp task
{
my_sleep (SLEEPTIME);
includedtids[included] = omp_get_thread_num();
} /* end of omp included task of the final task */
my_sleep (SLEEPTIME);
} /* end of if it is a final task*/
} /* end of omp task */
} /* end of for */
} /* end of single */
} /*end of parallel */
/* Now we ckeck if more than one thread executed the final task and its included task. */
for (i = 10; i < NUM_TASKS; i++) {
if (tids[i] != includedtids[i]) {
error++;
}
}
return (error==0);
} /* end of check_paralel_for_private */
int main()
{
int i;
int num_failed=0;
for(i = 0; i < REPETITIONS; i++) {
if(!test_omp_task_final()) {
num_failed++;
}
}
return num_failed;
}
| 775 |
777 | <gh_stars>100-1000
// Copyright 2016 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "core/animation/CSSImageSliceInterpolationType.h"
#include "core/animation/CSSLengthInterpolationType.h"
#include "core/animation/ImageSlicePropertyFunctions.h"
#include "core/css/CSSBorderImageSliceValue.h"
#include "core/css/resolver/StyleResolverState.h"
#include "wtf/PtrUtil.h"
#include <memory>
namespace blink {
namespace {
enum SideIndex : unsigned {
SideTop,
SideRight,
SideBottom,
SideLeft,
SideIndexCount,
};
struct SliceTypes {
explicit SliceTypes(const ImageSlice& slice) {
isNumber[SideTop] = slice.slices.top().isFixed();
isNumber[SideRight] = slice.slices.right().isFixed();
isNumber[SideBottom] = slice.slices.bottom().isFixed();
isNumber[SideLeft] = slice.slices.left().isFixed();
fill = slice.fill;
}
explicit SliceTypes(const CSSBorderImageSliceValue& slice) {
isNumber[SideTop] = slice.slices().top()->isPrimitiveValue() &&
toCSSPrimitiveValue(slice.slices().top())->isNumber();
isNumber[SideRight] =
slice.slices().right()->isPrimitiveValue() &&
toCSSPrimitiveValue(slice.slices().right())->isNumber();
isNumber[SideBottom] =
slice.slices().bottom()->isPrimitiveValue() &&
toCSSPrimitiveValue(slice.slices().bottom())->isNumber();
isNumber[SideLeft] = slice.slices().left()->isPrimitiveValue() &&
toCSSPrimitiveValue(slice.slices().left())->isNumber();
fill = slice.fill();
}
bool operator==(const SliceTypes& other) const {
for (size_t i = 0; i < SideIndexCount; i++) {
if (isNumber[i] != other.isNumber[i])
return false;
}
return fill == other.fill;
}
bool operator!=(const SliceTypes& other) const { return !(*this == other); }
// If a side is not a number then it is a percentage.
bool isNumber[SideIndexCount];
bool fill;
};
} // namespace
class CSSImageSliceNonInterpolableValue : public NonInterpolableValue {
public:
static PassRefPtr<CSSImageSliceNonInterpolableValue> create(
const SliceTypes& types) {
return adoptRef(new CSSImageSliceNonInterpolableValue(types));
}
const SliceTypes& types() const { return m_types; }
DECLARE_NON_INTERPOLABLE_VALUE_TYPE();
private:
CSSImageSliceNonInterpolableValue(const SliceTypes& types) : m_types(types) {}
const SliceTypes m_types;
};
DEFINE_NON_INTERPOLABLE_VALUE_TYPE(CSSImageSliceNonInterpolableValue);
DEFINE_NON_INTERPOLABLE_VALUE_TYPE_CASTS(CSSImageSliceNonInterpolableValue);
namespace {
class UnderlyingSliceTypesChecker
: public InterpolationType::ConversionChecker {
public:
static std::unique_ptr<UnderlyingSliceTypesChecker> create(
const SliceTypes& underlyingTypes) {
return WTF::wrapUnique(new UnderlyingSliceTypesChecker(underlyingTypes));
}
static SliceTypes getUnderlyingSliceTypes(
const InterpolationValue& underlying) {
return toCSSImageSliceNonInterpolableValue(*underlying.nonInterpolableValue)
.types();
}
private:
UnderlyingSliceTypesChecker(const SliceTypes& underlyingTypes)
: m_underlyingTypes(underlyingTypes) {}
bool isValid(const InterpolationEnvironment&,
const InterpolationValue& underlying) const final {
return m_underlyingTypes == getUnderlyingSliceTypes(underlying);
}
const SliceTypes m_underlyingTypes;
};
class InheritedSliceTypesChecker : public InterpolationType::ConversionChecker {
public:
static std::unique_ptr<InheritedSliceTypesChecker> create(
CSSPropertyID property,
const SliceTypes& inheritedTypes) {
return WTF::wrapUnique(
new InheritedSliceTypesChecker(property, inheritedTypes));
}
private:
InheritedSliceTypesChecker(CSSPropertyID property,
const SliceTypes& inheritedTypes)
: m_property(property), m_inheritedTypes(inheritedTypes) {}
bool isValid(const InterpolationEnvironment& environment,
const InterpolationValue& underlying) const final {
return m_inheritedTypes ==
SliceTypes(ImageSlicePropertyFunctions::getImageSlice(
m_property, *environment.state().parentStyle()));
}
const CSSPropertyID m_property;
const SliceTypes m_inheritedTypes;
};
InterpolationValue convertImageSlice(const ImageSlice& slice, double zoom) {
std::unique_ptr<InterpolableList> list =
InterpolableList::create(SideIndexCount);
const Length* sides[SideIndexCount] = {};
sides[SideTop] = &slice.slices.top();
sides[SideRight] = &slice.slices.right();
sides[SideBottom] = &slice.slices.bottom();
sides[SideLeft] = &slice.slices.left();
for (size_t i = 0; i < SideIndexCount; i++) {
const Length& side = *sides[i];
list->set(i, InterpolableNumber::create(
side.isFixed() ? side.pixels() / zoom : side.percent()));
}
return InterpolationValue(
std::move(list),
CSSImageSliceNonInterpolableValue::create(SliceTypes(slice)));
}
} // namespace
InterpolationValue CSSImageSliceInterpolationType::maybeConvertNeutral(
const InterpolationValue& underlying,
ConversionCheckers& conversionCheckers) const {
SliceTypes underlyingTypes =
UnderlyingSliceTypesChecker::getUnderlyingSliceTypes(underlying);
conversionCheckers.push_back(
UnderlyingSliceTypesChecker::create(underlyingTypes));
LengthBox zeroBox(
Length(0, underlyingTypes.isNumber[SideTop] ? Fixed : Percent),
Length(0, underlyingTypes.isNumber[SideRight] ? Fixed : Percent),
Length(0, underlyingTypes.isNumber[SideBottom] ? Fixed : Percent),
Length(0, underlyingTypes.isNumber[SideLeft] ? Fixed : Percent));
return convertImageSlice(ImageSlice(zeroBox, underlyingTypes.fill), 1);
}
InterpolationValue CSSImageSliceInterpolationType::maybeConvertInitial(
const StyleResolverState&,
ConversionCheckers& conversionCheckers) const {
return convertImageSlice(
ImageSlicePropertyFunctions::getInitialImageSlice(cssProperty()), 1);
}
InterpolationValue CSSImageSliceInterpolationType::maybeConvertInherit(
const StyleResolverState& state,
ConversionCheckers& conversionCheckers) const {
const ImageSlice& inheritedImageSlice =
ImageSlicePropertyFunctions::getImageSlice(cssProperty(),
*state.parentStyle());
conversionCheckers.push_back(InheritedSliceTypesChecker::create(
cssProperty(), SliceTypes(inheritedImageSlice)));
return convertImageSlice(inheritedImageSlice,
state.parentStyle()->effectiveZoom());
}
InterpolationValue CSSImageSliceInterpolationType::maybeConvertValue(
const CSSValue& value,
const StyleResolverState&,
ConversionCheckers&) const {
if (!value.isBorderImageSliceValue())
return nullptr;
const CSSBorderImageSliceValue& slice = toCSSBorderImageSliceValue(value);
std::unique_ptr<InterpolableList> list =
InterpolableList::create(SideIndexCount);
const CSSValue* sides[SideIndexCount];
sides[SideTop] = slice.slices().top();
sides[SideRight] = slice.slices().right();
sides[SideBottom] = slice.slices().bottom();
sides[SideLeft] = slice.slices().left();
for (size_t i = 0; i < SideIndexCount; i++) {
const CSSPrimitiveValue& side = *toCSSPrimitiveValue(sides[i]);
DCHECK(side.isNumber() || side.isPercentage());
list->set(i, InterpolableNumber::create(side.getDoubleValue()));
}
return InterpolationValue(
std::move(list),
CSSImageSliceNonInterpolableValue::create(SliceTypes(slice)));
}
InterpolationValue
CSSImageSliceInterpolationType::maybeConvertStandardPropertyUnderlyingValue(
const StyleResolverState& state) const {
const ComputedStyle& style = *state.style();
return convertImageSlice(
ImageSlicePropertyFunctions::getImageSlice(cssProperty(), style),
style.effectiveZoom());
}
PairwiseInterpolationValue CSSImageSliceInterpolationType::maybeMergeSingles(
InterpolationValue&& start,
InterpolationValue&& end) const {
const SliceTypes& startSliceTypes =
toCSSImageSliceNonInterpolableValue(*start.nonInterpolableValue).types();
const SliceTypes& endSliceTypes =
toCSSImageSliceNonInterpolableValue(*end.nonInterpolableValue).types();
if (startSliceTypes != endSliceTypes)
return nullptr;
return PairwiseInterpolationValue(std::move(start.interpolableValue),
std::move(end.interpolableValue),
std::move(start.nonInterpolableValue));
}
void CSSImageSliceInterpolationType::composite(
UnderlyingValueOwner& underlyingValueOwner,
double underlyingFraction,
const InterpolationValue& value,
double interpolationFraction) const {
const SliceTypes& underlyingTypes =
toCSSImageSliceNonInterpolableValue(
*underlyingValueOwner.value().nonInterpolableValue)
.types();
const SliceTypes& types =
toCSSImageSliceNonInterpolableValue(*value.nonInterpolableValue).types();
if (underlyingTypes == types)
underlyingValueOwner.mutableValue().interpolableValue->scaleAndAdd(
underlyingFraction, *value.interpolableValue);
else
underlyingValueOwner.set(*this, value);
}
void CSSImageSliceInterpolationType::applyStandardPropertyValue(
const InterpolableValue& interpolableValue,
const NonInterpolableValue* nonInterpolableValue,
StyleResolverState& state) const {
ComputedStyle& style = *state.style();
const InterpolableList& list = toInterpolableList(interpolableValue);
const SliceTypes& types =
toCSSImageSliceNonInterpolableValue(nonInterpolableValue)->types();
const auto& convertSide = [&types, &list, &style](size_t index) {
float value =
clampTo<float>(toInterpolableNumber(list.get(index))->value(), 0);
return types.isNumber[index] ? Length(value * style.effectiveZoom(), Fixed)
: Length(value, Percent);
};
LengthBox box(convertSide(SideTop), convertSide(SideRight),
convertSide(SideBottom), convertSide(SideLeft));
ImageSlicePropertyFunctions::setImageSlice(cssProperty(), style,
ImageSlice(box, types.fill));
}
} // namespace blink
| 3,745 |
2,350 | #pragma once
#include "UnrealEnginePython.h"
#include "Runtime/Online/Voice/Public/VoiceModule.h"
#include "Runtime/Online/Voice/Public/Interfaces/VoiceCapture.h"
typedef struct {
PyObject_HEAD
/* Type-specific fields go here. */
TSharedRef<IVoiceCapture> voice_capture;
} ue_PyIVoiceCapture;
void ue_python_init_ivoice_capture(PyObject *);
| 126 |
2,937 | <reponame>fegorsch/ogre
/*
-----------------------------------------------------------------------------
This source file is part of OGRE
(Object-oriented Graphics Rendering Engine)
For the latest info, see http://www.ogre3d.org/
Copyright (c) 2000-2014 Torus <NAME> Ltd
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.
-----------------------------------------------------------------------------
*/
#include "OgreGLVertexArrayObject.h"
#include "OgreRoot.h"
#include "OgreLogManager.h"
#include "OgreGLSLProgramCommon.h"
#include "OgreGLRenderSystemCommon.h"
namespace Ogre {
GLVertexArrayObject::GLVertexArrayObject() : mCreatorContext(0), mVAO(0), mNeedsUpdate(true), mVertexStart(0) {
}
GLVertexArrayObject::~GLVertexArrayObject()
{
if(mVAO != 0)
{
GLRenderSystemCommon* rs = static_cast<GLRenderSystemCommon*>(Root::getSingleton().getRenderSystem());
rs->_destroyVao(mCreatorContext, mVAO);
}
}
void GLVertexArrayObject::bind(GLRenderSystemCommon* rs)
{
if(mCreatorContext && mCreatorContext != rs->_getCurrentContext()) // VAO is unusable with current context, destroy it
{
if(mVAO != 0)
rs->_destroyVao(mCreatorContext, mVAO);
mCreatorContext = 0;
mVAO = 0;
mNeedsUpdate = true;
}
if(!mCreatorContext && rs->getCapabilities()->hasCapability(RSC_VAO)) // create VAO lazy or recreate after destruction
{
mCreatorContext = rs->_getCurrentContext();
mVAO = rs->_createVao();
mNeedsUpdate = true;
}
rs->_bindVao(mCreatorContext, mVAO);
}
bool GLVertexArrayObject::needsUpdate(VertexBufferBinding* vertexBufferBinding,
size_t vertexStart)
{
if(mNeedsUpdate)
return true;
VertexDeclaration::VertexElementList::const_iterator elemIter, elemEnd;
elemEnd = mElementList.end();
for (elemIter = mElementList.begin(); elemIter != elemEnd; ++elemIter)
{
const VertexElement & elem = *elemIter;
uint16 source = elem.getSource();
if (!vertexBufferBinding->isBufferBound(source))
continue; // Skip unbound elements
VertexElementSemantic sem = elem.getSemantic();
unsigned short elemIndex = elem.getIndex();
uint32 attrib = (uint32)GLSLProgramCommon::getFixedAttributeIndex(sem, elemIndex);
const HardwareVertexBufferSharedPtr& vertexBuffer = vertexBufferBinding->getBuffer(source);
if (std::find(mAttribsBound.begin(), mAttribsBound.end(),
std::make_pair(attrib, vertexBuffer.get())) == mAttribsBound.end())
return true;
if (vertexBuffer->isInstanceData() &&
std::find(mInstanceAttribsBound.begin(), mInstanceAttribsBound.end(), attrib) ==
mInstanceAttribsBound.end())
return true;
}
if(vertexStart != mVertexStart) {
return true;
}
return false;
}
void GLVertexArrayObject::bindToGpu(GLRenderSystemCommon* rs,
VertexBufferBinding* vertexBufferBinding,
size_t vertexStart)
{
mAttribsBound.clear();
mInstanceAttribsBound.clear();
VertexDeclaration::VertexElementList::const_iterator elemIter, elemEnd;
elemEnd = mElementList.end();
for (elemIter = mElementList.begin(); elemIter != elemEnd; ++elemIter)
{
const VertexElement& elem = *elemIter;
uint16 source = elem.getSource();
if (!vertexBufferBinding->isBufferBound(source))
continue; // Skip unbound elements
VertexElementSemantic sem = elem.getSemantic();
unsigned short elemIndex = elem.getIndex();
uint32 attrib = (uint32)GLSLProgramCommon::getFixedAttributeIndex(sem, elemIndex);
const HardwareVertexBufferSharedPtr& vertexBuffer = vertexBufferBinding->getBuffer(source);
mAttribsBound.push_back(std::make_pair(attrib, vertexBuffer.get()));
rs->bindVertexElementToGpu(elem, vertexBuffer, vertexStart);
if (vertexBuffer->isInstanceData())
mInstanceAttribsBound.push_back(attrib);
}
mVertexStart = vertexStart;
mNeedsUpdate = false;
}
}
| 2,274 |
1,515 | <filename>qigsaw-android/splitinstaller/src/main/java/com/iqiyi/android/qigsaw/core/splitinstall/SplitSessionInstallerImpl.java
/*
* MIT License
*
* Copyright (c) 2019-present, iQIYI, Inc. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a copy
* of this software and associated documentation files (the "Software"), to deal
* in the Software without restriction, including without limitation the rights
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
* copies of the Software, and to permit persons to whom the Software is
* furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in all
* copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
package com.iqiyi.android.qigsaw.core.splitinstall;
import com.iqiyi.android.qigsaw.core.splitrequest.splitinfo.SplitInfo;
import java.util.List;
import java.util.concurrent.Executor;
final class SplitSessionInstallerImpl implements SplitSessionInstaller {
private final Executor executor;
private final SplitInstallSessionManager sessionManager;
private final SplitInstaller splitInstaller;
SplitSessionInstallerImpl(SplitInstaller splitInstaller, SplitInstallSessionManager sessionManager, Executor executor) {
this.splitInstaller = splitInstaller;
this.sessionManager = sessionManager;
this.executor = executor;
}
@Override
public void install(int sessionId, List<SplitInfo> splitInfoList) {
executor.execute(new SplitStartInstallTask(sessionId, splitInstaller, sessionManager, splitInfoList));
}
}
| 617 |
3,680 | //
// Copyright 2016 Pixar
//
// Licensed under the Apache License, Version 2.0 (the "Apache License")
// with the following modification; you may not use this file except in
// compliance with the Apache License and the following modification to it:
// Section 6. Trademarks. is deleted and replaced with:
//
// 6. Trademarks. This License does not grant permission to use the trade
// names, trademarks, service marks, or product names of the Licensor
// and its affiliates, except as required to comply with Section 4(c) of
// the License and to reproduce the content of the NOTICE file.
//
// You may obtain a copy of the Apache License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the Apache License with the above modification is
// distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the Apache License for the specific
// language governing permissions and limitations under the Apache License.
//
#include "pxr/usd/usdRender/tokens.h"
PXR_NAMESPACE_OPEN_SCOPE
UsdRenderTokensType::UsdRenderTokensType() :
adjustApertureHeight("adjustApertureHeight", TfToken::Immortal),
adjustApertureWidth("adjustApertureWidth", TfToken::Immortal),
adjustPixelAspectRatio("adjustPixelAspectRatio", TfToken::Immortal),
aspectRatioConformPolicy("aspectRatioConformPolicy", TfToken::Immortal),
camera("camera", TfToken::Immortal),
color3f("color3f", TfToken::Immortal),
cropAperture("cropAperture", TfToken::Immortal),
dataType("dataType", TfToken::Immortal),
dataWindowNDC("dataWindowNDC", TfToken::Immortal),
expandAperture("expandAperture", TfToken::Immortal),
full("full", TfToken::Immortal),
includedPurposes("includedPurposes", TfToken::Immortal),
instantaneousShutter("instantaneousShutter", TfToken::Immortal),
intrinsic("intrinsic", TfToken::Immortal),
lpe("lpe", TfToken::Immortal),
materialBindingPurposes("materialBindingPurposes", TfToken::Immortal),
orderedVars("orderedVars", TfToken::Immortal),
pixelAspectRatio("pixelAspectRatio", TfToken::Immortal),
preview("preview", TfToken::Immortal),
primvar("primvar", TfToken::Immortal),
productName("productName", TfToken::Immortal),
products("products", TfToken::Immortal),
productType("productType", TfToken::Immortal),
raster("raster", TfToken::Immortal),
raw("raw", TfToken::Immortal),
renderSettingsPrimPath("renderSettingsPrimPath", TfToken::Immortal),
resolution("resolution", TfToken::Immortal),
sourceName("sourceName", TfToken::Immortal),
sourceType("sourceType", TfToken::Immortal),
allTokens({
adjustApertureHeight,
adjustApertureWidth,
adjustPixelAspectRatio,
aspectRatioConformPolicy,
camera,
color3f,
cropAperture,
dataType,
dataWindowNDC,
expandAperture,
full,
includedPurposes,
instantaneousShutter,
intrinsic,
lpe,
materialBindingPurposes,
orderedVars,
pixelAspectRatio,
preview,
primvar,
productName,
products,
productType,
raster,
raw,
renderSettingsPrimPath,
resolution,
sourceName,
sourceType
})
{
}
TfStaticData<UsdRenderTokensType> UsdRenderTokens;
PXR_NAMESPACE_CLOSE_SCOPE
| 1,252 |
314 | package com.lynnchurch.pulltorefresh;
import android.content.Context;
import android.support.v7.widget.RecyclerView;
import android.view.LayoutInflater;
import android.view.View;
import android.view.ViewGroup;
import android.widget.TextView;
import com.jingchen.pulltorefresh.WrapRecyclerView;
import com.lsjwzh.widget.materialloadingprogressbar.CircleProgressBar;
import java.util.ArrayList;
/**
* 自动加载更多适配器
*/
public abstract class BaseAutoLoadMoreAdapter<T> extends RecyclerView.Adapter<RecyclerView.ViewHolder>
{
private static final String TAG = BaseAutoLoadMoreAdapter.class.getSimpleName();
protected Context mContext;
protected ArrayList<T> mData;
private OnItemClickListener mOnItemClickListener;
private OnLoadmoreListener mOnLoadmoreListener;
private int mLastPosition; // 正常项最后一项的位置
private WrapRecyclerView mParentView;
private int mParentHeight; // item父视图的高度
private int mItemsHeight; // 所有item的总高度
private TextView tv_hint;
private CircleProgressBar progressBar;
public BaseAutoLoadMoreAdapter(Context context, WrapRecyclerView recyclerView, ArrayList<T> data)
{
mContext = context;
mParentView = recyclerView;
initFooter(recyclerView);
mData = data;
mLastPosition = mData.size() - 1;
}
private void initFooter(WrapRecyclerView recyclerView)
{
View footer = LayoutInflater.from(mContext).inflate(R.layout.loadmore, null);
tv_hint = (TextView) footer.findViewById(R.id.tv_hint);
progressBar = (CircleProgressBar) footer.findViewById(R.id.progressBar);
recyclerView.addFootView(footer);
showLoading(false);
}
@Override
final public RecyclerView.ViewHolder onCreateViewHolder(ViewGroup parent, int viewType)
{
mParentHeight = parent.getHeight();
return onCreateBaseViewHolder(parent, viewType);
}
public abstract BaseViewHolder onCreateBaseViewHolder(ViewGroup parent, int viewType);
public abstract int getItemHeight(RecyclerView.ViewHolder holder);
public void resetItemHeight()
{
mItemsHeight = 0;
}
@Override
final public void onBindViewHolder(RecyclerView.ViewHolder holder, final int position)
{
if(!isOverParent())
{
showLoading(false);
mItemsHeight+=getItemHeight(holder);
}
if (holder instanceof BaseViewHolder)
{
final BaseViewHolder viewHolder = (BaseViewHolder) holder;
if (null != mOnItemClickListener)
{
viewHolder.itemView.setOnClickListener(new View.OnClickListener()
{
@Override
public void onClick(View v)
{
mOnItemClickListener.onItemClick(viewHolder.itemView, position);
}
});
viewHolder.itemView.setOnLongClickListener(new View.OnLongClickListener()
{
@Override
public boolean onLongClick(View v)
{
mOnItemClickListener.onItemLongClick(viewHolder.itemView, position);
return true;
}
});
}
onBindBaseViewHolder(viewHolder, position);
}
if (position + 1 > mLastPosition && null != mOnLoadmoreListener && isOverParent())
{
showLoading(true);
mOnLoadmoreListener.onLoadmore();
mLastPosition = mData.size();
}
}
public abstract void onBindBaseViewHolder(BaseViewHolder holder, final int position);
/**
* item是否撑满父视图
*
* @return
*/
public boolean isOverParent()
{
return mParentHeight < mItemsHeight + mParentView.getHeaderHeight() ? true : false;
}
@Override
public int getItemCount()
{
return mData.size();
}
public static class BaseViewHolder extends RecyclerView.ViewHolder
{
View itemView;
public BaseViewHolder(View v)
{
super(v);
itemView = v;
}
}
/**
* 当加载失败
*/
public void onFailed()
{
progressBar.setVisibility(View.GONE);
tv_hint.setVisibility(View.VISIBLE);
tv_hint.setText("点击重试");
tv_hint.setOnClickListener(new View.OnClickListener()
{
@Override
public void onClick(View v)
{
if (null != mOnLoadmoreListener)
{
progressBar.setVisibility(View.VISIBLE);
tv_hint.setVisibility(View.GONE);
mOnLoadmoreListener.onLoadmore();
}
}
});
}
/**
* 当没有更多
*/
public void onNothing()
{
progressBar.setVisibility(View.GONE);
tv_hint.setVisibility(View.VISIBLE);
tv_hint.setText("没有更多");
tv_hint.setOnClickListener(null);
}
/**
* 显示加载视图
*
* @param show
*/
private void showLoading(boolean show)
{
if (show)
{
progressBar.setVisibility(View.VISIBLE);
} else
{
progressBar.setVisibility(View.GONE);
}
}
/**
* 当数据改变时调用,替换notifyDataSetChanged()
*/
public void notifyDataChanged()
{
notifyDataSetChanged();
mParentView.requestLayout();
}
/**
* 设置Item监听器
*
* @param onItemClickListener
*/
public void setOnItemClickListener(OnItemClickListener onItemClickListener)
{
mOnItemClickListener = onItemClickListener;
}
/**
* Item点击监听
*/
public interface OnItemClickListener
{
void onItemClick(View view, int position);
void onItemLongClick(View view, int position);
}
/**
* 设置加载更多监听
*
* @param listener
*/
public void setOnLoadmoreListener(OnLoadmoreListener listener)
{
mOnLoadmoreListener = listener;
}
/**
* 加载更多监听
*/
public interface OnLoadmoreListener
{
void onLoadmore();
}
} | 3,030 |
1,241 | import math
import os
from time import time
from typing import List, Dict, Union, Tuple
from tensorflow.python.keras.models import load_model
from donkeycar.config import Config
from donkeycar.parts.keras import KerasPilot
from donkeycar.parts.interpreter import keras_model_to_tflite, \
saved_model_to_tensor_rt
from donkeycar.pipeline.database import PilotDatabase
from donkeycar.pipeline.sequence import TubRecord, TubSequence, TfmIterator
from donkeycar.pipeline.types import TubDataset
from donkeycar.pipeline.augmentations import ImageAugmentation
from donkeycar.utils import get_model_by_type, normalize_image, train_test_split
import tensorflow as tf
import numpy as np
class BatchSequence(object):
"""
The idea is to have a shallow sequence with types that can hydrate
themselves to np.ndarray initially and later into the types required by
tf.data (i.e. dictionaries or np.ndarrays).
"""
def __init__(self,
model: KerasPilot,
config: Config,
records: List[TubRecord],
is_train: bool) -> None:
self.model = model
self.config = config
self.sequence = TubSequence(records)
self.batch_size = self.config.BATCH_SIZE
self.is_train = is_train
self.augmentation = ImageAugmentation(config, 'AUGMENTATIONS')
self.transformation = ImageAugmentation(config, 'TRANSFORMATIONS')
self.pipeline = self._create_pipeline()
def __len__(self) -> int:
return math.ceil(len(self.pipeline) / self.batch_size)
def image_processor(self, img_arr):
""" Transformes the images and augments if in training. Then
normalizes it. """
img_arr = self.transformation.run(img_arr)
if self.is_train:
img_arr = self.augmentation.run(img_arr)
norm_img = normalize_image(img_arr)
return norm_img
def _create_pipeline(self) -> TfmIterator:
""" This can be overridden if more complicated pipelines are
required """
# 1. Initialise TubRecord -> x, y transformations
def get_x(record: TubRecord) -> Dict[str, Union[float, np.ndarray]]:
""" Extracting x from record for training"""
out_tuple = self.model.x_transform_and_process(
record, self.image_processor)
# convert tuple to dictionary which is understood by tf.data
out_dict = self.model.x_translate(out_tuple)
return out_dict
def get_y(record: TubRecord) -> Dict[str, Union[float, np.ndarray]]:
""" Extracting y from record for training """
y0 = self.model.y_transform(record)
y1 = self.model.y_translate(y0)
return y1
# 2. Build pipeline using the transformations
pipeline = self.sequence.build_pipeline(x_transform=get_x,
y_transform=get_y)
return pipeline
def create_tf_data(self) -> tf.data.Dataset:
""" Assembles the tf data pipeline """
dataset = tf.data.Dataset.from_generator(
generator=lambda: self.pipeline,
output_types=self.model.output_types(),
output_shapes=self.model.output_shapes())
return dataset.repeat().batch(self.batch_size)
def get_model_train_details(database: PilotDatabase, model: str = None) \
-> Tuple[str, int]:
if not model:
model_name, model_num = database.generate_model_name()
else:
model_name, model_num = os.path.abspath(model), 0
return model_name, model_num
def train(cfg: Config, tub_paths: str, model: str = None,
model_type: str = None, transfer: str = None, comment: str = None) \
-> tf.keras.callbacks.History:
"""
Train the model
"""
database = PilotDatabase(cfg)
if model_type is None:
model_type = cfg.DEFAULT_MODEL_TYPE
model_path, model_num = \
get_model_train_details(database, model)
base_path = os.path.splitext(model_path)[0]
kl = get_model_by_type(model_type, cfg)
if transfer:
kl.load(transfer)
if cfg.PRINT_MODEL_SUMMARY:
print(kl.interpreter.model.summary())
tubs = tub_paths.split(',')
all_tub_paths = [os.path.expanduser(tub) for tub in tubs]
dataset = TubDataset(config=cfg, tub_paths=all_tub_paths,
seq_size=kl.seq_size())
training_records, validation_records \
= train_test_split(dataset.get_records(), shuffle=True,
test_size=(1. - cfg.TRAIN_TEST_SPLIT))
print(f'Records # Training {len(training_records)}')
print(f'Records # Validation {len(validation_records)}')
# We need augmentation in validation when using crop / trapeze
training_pipe = BatchSequence(kl, cfg, training_records, is_train=True)
validation_pipe = BatchSequence(kl, cfg, validation_records, is_train=False)
tune = tf.data.experimental.AUTOTUNE
dataset_train = training_pipe.create_tf_data().prefetch(tune)
dataset_validate = validation_pipe.create_tf_data().prefetch(tune)
train_size = len(training_pipe)
val_size = len(validation_pipe)
assert val_size > 0, "Not enough validation data, decrease the batch " \
"size or add more data."
history = kl.train(model_path=model_path,
train_data=dataset_train,
train_steps=train_size,
batch_size=cfg.BATCH_SIZE,
validation_data=dataset_validate,
validation_steps=val_size,
epochs=cfg.MAX_EPOCHS,
verbose=cfg.VERBOSE_TRAIN,
min_delta=cfg.MIN_DELTA,
patience=cfg.EARLY_STOP_PATIENCE,
show_plot=cfg.SHOW_PLOT)
if getattr(cfg, 'CREATE_TF_LITE', True):
tf_lite_model_path = f'{base_path}.tflite'
keras_model_to_tflite(model_path, tf_lite_model_path)
if getattr(cfg, 'CREATE_TENSOR_RT', False):
# load h5 (ie. keras) model
model_rt = load_model(model_path)
# save in tensorflow savedmodel format (i.e. directory)
model_rt.save(f'{base_path}.savedmodel')
# pass savedmodel to the rt converter
saved_model_to_tensor_rt(f'{base_path}.savedmodel', f'{base_path}.trt')
database_entry = {
'Number': model_num,
'Name': os.path.basename(base_path),
'Type': str(kl),
'Tubs': tub_paths,
'Time': time(),
'History': history.history,
'Transfer': os.path.basename(transfer) if transfer else None,
'Comment': comment,
'Config': str(cfg)
}
database.add_entry(database_entry)
database.write()
return history
| 3,030 |
435 | <filename>pycon-tw-2021/videos/supercharging-your-jupyter-and-ml-workflow-with-primehub-chia-liang-kao-pycon-taiwan-2021.json<gh_stars>100-1000
{
"description": "Title: Supercharging your Jupyter and Machine Learning Workflow with PrimeHub \u2013 <NAME> \n\nDay 2, 10:35-11:20\n\nAbstract\n\nJupyter has become a critical component of the machine learning life cycle. However scaled enterprise deployments and making the Data Science Experience frictionless remain challenging. We address a few common issues with PrimeHub, an open-source enterprise offering based on JupyterHub, with the capability to work with popular ML tooling like mlflow, labelstudio, and streamlit. This talk also investigates the MLOps trends adjacent to the Jupyter ecosystem.\n\nDescription\n\nThis talk is intended for audience interested in larger scale Jupyter environment deployment in their organisation, particularly for machine learning applications.\n\nPrimeHub is an open-source enterprise offering based on JupyterHub, which also allows orchestrating tools like mlflow, labelstudio, and streamlit, to assemble your own end-to-end ML toolbox for your team. We address a few common hurdles:\n\nadvanced resource scheduling for heterogenous GPU clusters\nmulti-tenancy and project isolations\ncustomizable and consistent ML environments per project\nmanaging jupyterlab extensions for teams\nhybrid development environment such as vscode and ssh within notebook instances\nauthorization and data access management\nusage reporting: allocation and utilization\nWe also investigate a few trends adjacent to the day-to-day jupyter environment used by data scientists and data engineers, where the roles become more cross functional in the age of MLOps:\n\nWorking with job schedulers from within Notebook\nManaging ML model deployments, CI/CD integration\nModel monitoring and retraining\n\nSlides not uploaded by the speaker.\nHackMD: https://hackmd.io/@pycontw/2021/%2F%40pycontw%2Frk-phz9zF\n\nSpeaker: <NAME>\n\nclkao (<NAME>) has been an open source software developer since 2000. He believes that good collaboration model and tools drive innovation. In 2013, he created SVK, a distributed version control system that helps developers collaborate. He co-founded the g0v.tw community in 2012, advocating information transparency and digital-activism through open source model. g0v.tw was awarded as \"Digital Communities: Award of Distinction\" by Prix Ars Electronica 2018. He started InfuseAI in 2018 to enable data scientists to thrive, and to help wider adoption of AI across industries.",
"speakers": [
"<NAME>"
],
"recorded": "2021-10-03",
"title": "Supercharging your Jupyter and ML Workflow with PrimeHub \u2013 Chia-liang Kao (PyCon Taiwan 2021)",
"thumbnail_url": "https://i.ytimg.com/vi/qf0U1JIgCFY/hqdefault.jpg",
"videos": [
{
"type": "youtube",
"url": "https://www.youtube.com/watch?v=qf0U1JIgCFY"
}
]
} | 869 |
892 | {
"schema_version": "1.2.0",
"id": "GHSA-3vv6-22qm-hgwq",
"modified": "2022-05-13T01:36:56Z",
"published": "2022-05-13T01:36:56Z",
"aliases": [
"CVE-2017-2348"
],
"details": "The Juniper Enhanced jdhcpd daemon may experience high CPU utilization, or crash and restart upon receipt of an invalid IPv6 UDP packet. Both high CPU utilization and repeated crashes of the jdhcpd daemon can result in a denial of service as DHCP service is interrupted. No other Juniper Networks products or platforms are affected by this issue. Affected releases are Juniper Networks Junos OS 14.1X53 prior to 14.1X53-D12, 14.1X53-D38, 14.1X53-D40 on QFX, EX, QFabric System; 15.1 prior to 15.1F2-S18, 15.1R4 on all products and platforms; 15.1X49 prior to 15.1X49-D80 on SRX; 15.1X53 prior to 15.1X53-D51, 15.1X53-D60 on NFX, QFX, EX.",
"severity": [
{
"type": "CVSS_V3",
"score": "CVSS:3.0/AV:N/AC:L/PR:N/UI:N/S:U/C:N/I:N/A:H"
}
],
"affected": [
],
"references": [
{
"type": "ADVISORY",
"url": "https://nvd.nist.gov/vuln/detail/CVE-2017-2348"
},
{
"type": "WEB",
"url": "https://kb.juniper.net/JSA10800"
},
{
"type": "WEB",
"url": "http://www.securitytracker.com/id/1038899"
}
],
"database_specific": {
"cwe_ids": [
"CWE-400"
],
"severity": "HIGH",
"github_reviewed": false
}
} | 627 |
325 | from .process import Processing, Mask, F
from .partition import Partition, TimePartition | 21 |
672 | <filename>xnu-4903.241.1/bsd/arm/disklabel.h
/*
* Copyright (c) 2000-2007 Apple Inc. All rights reserved.
*/
#ifndef _MACHINE_DISKLABEL_H_
#define _MACHINE_DISKLABEL_H_
#include <sys/appleapiopts.h>
#ifdef __APPLE_API_OBSOLETE
#define LABELSECTOR (1024 / DEV_BSIZE) /* sector containing label */
#define LABELOFFSET 0 /* offset of label in sector */
#define MAXPARTITIONS 8 /* number of partitions */
#define RAW_PART 2 /* raw partition: xx?c */
/* Just a dummy */
struct cpu_disklabel {
int cd_dummy; /* must have one element. */
};
#endif /* __APPLE_API_OBSOLETE */
#endif /* _MACHINE_DISKLABEL_H_ */
| 256 |
398 | <gh_stars>100-1000
#!/usr/bin/env python3
import argparse, csv, sys, pickle, collections, time
from common import *
if len(sys.argv) == 1:
sys.argv.append('-h')
parser = argparse.ArgumentParser()
parser.add_argument('tr_src_path', type=str)
parser.add_argument('va_src_path', type=str)
parser.add_argument('tr_app_dst_path', type=str)
parser.add_argument('va_app_dst_path', type=str)
parser.add_argument('tr_site_dst_path', type=str)
parser.add_argument('va_site_dst_path', type=str)
args = vars(parser.parse_args())
FIELDS = ['id','click','hour','banner_pos','device_id','device_ip','device_model','device_conn_type','C14','C17','C20','C21']
NEW_FIELDS = FIELDS+['pub_id','pub_domain','pub_category','device_id_count','device_ip_count','user_count','smooth_user_hour_count','user_click_histroy']
id_cnt = collections.defaultdict(int)
ip_cnt = collections.defaultdict(int)
user_cnt = collections.defaultdict(int)
user_hour_cnt = collections.defaultdict(int)
start = time.time()
def scan(path):
for i, row in enumerate(csv.DictReader(open(path)), start=1):
if i % 1000000 == 0:
sys.stderr.write('{0:6.0f} {1}m\n'.format(time.time()-start,int(i/1000000)))
user = def_user(row)
id_cnt[row['device_id']] += 1
ip_cnt[row['device_ip']] += 1
user_cnt[user] += 1
user_hour_cnt[user+'-'+row['hour']] += 1
history = collections.defaultdict(lambda: {'history': '', 'buffer': '', 'prev_hour': ''})
def gen_data(src_path, dst_app_path, dst_site_path, is_train):
reader = csv.DictReader(open(src_path))
writer_app = csv.DictWriter(open(dst_app_path, 'w'), NEW_FIELDS)
writer_site = csv.DictWriter(open(dst_site_path, 'w'), NEW_FIELDS)
writer_app.writeheader()
writer_site.writeheader()
for i, row in enumerate(reader, start=1):
if i % 1000000 == 0:
sys.stderr.write('{0:6.0f} {1}m\n'.format(time.time()-start,int(i/1000000)))
new_row = {}
for field in FIELDS:
new_row[field] = row[field]
new_row['device_id_count'] = id_cnt[row['device_id']]
new_row['device_ip_count'] = ip_cnt[row['device_ip']]
user, hour = def_user(row), row['hour']
new_row['user_count'] = user_cnt[user]
new_row['smooth_user_hour_count'] = str(user_hour_cnt[user+'-'+hour])
if has_id_info(row):
if history[user]['prev_hour'] != row['hour']:
history[user]['history'] = (history[user]['history'] + history[user]['buffer'])[-4:]
history[user]['buffer'] = ''
history[user]['prev_hour'] = row['hour']
new_row['user_click_histroy'] = history[user]['history']
if is_train:
history[user]['buffer'] += row['click']
else:
new_row['user_click_histroy'] = ''
if is_app(row):
new_row['pub_id'] = row['app_id']
new_row['pub_domain'] = row['app_domain']
new_row['pub_category'] = row['app_category']
writer_app.writerow(new_row)
else:
new_row['pub_id'] = row['site_id']
new_row['pub_domain'] = row['site_domain']
new_row['pub_category'] = row['site_category']
writer_site.writerow(new_row)
scan(args['tr_src_path'])
scan(args['va_src_path'])
print('======================scan complete======================')
gen_data(args['tr_src_path'], args['tr_app_dst_path'], args['tr_site_dst_path'], True)
gen_data(args['va_src_path'], args['va_app_dst_path'], args['va_site_dst_path'], False)
| 1,643 |
2,389 | /*
Copyright 2020 The Kubernetes Authors.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package io.kubernetes.client.informer.cache;
import java.util.List;
public interface Store<ApiType> {
/**
* add inserts an item into the store.
*
* @param obj specific obj
*/
void add(ApiType obj);
/**
* update sets an item in the store to its updated state.
*
* @param obj specific obj
*/
void update(ApiType obj);
/**
* delete removes an item from the store.
*
* @param obj specific obj
*/
void delete(ApiType obj);
/**
* Replace will delete the contents of 'c', using instead the given list.
*
* @param list list of objects
* @param resourceVersion specific resource version
*/
void replace(List<ApiType> list, String resourceVersion);
/** resync will send a resync event for each item. */
void resync();
/**
* listKeys returns a list of all the keys of the object currently in the store.
*
* @return list of all keys
*/
List<String> listKeys();
/**
* get returns the requested item.
*
* @param obj specific obj
* @return the requested item if exist
*/
Object get(ApiType obj);
/**
* getByKey returns the request item with specific key.
*
* @param key specific key
* @return the request item
*/
ApiType getByKey(String key);
/**
* list returns a list of all the items.
*
* @return list of all the items
*/
List<ApiType> list();
}
| 603 |
711 | <reponame>jingetiema2100/MicroCommunity<filename>service-store/src/main/java/com/java110/store/bmo/contractType/IDeleteContractTypeBMO.java<gh_stars>100-1000
package com.java110.store.bmo.contractType;
import com.java110.po.contractType.ContractTypePo;
import org.springframework.http.ResponseEntity;
public interface IDeleteContractTypeBMO {
/**
* 修改合同类型
* add by wuxw
* @param contractTypePo
* @return
*/
ResponseEntity<String> delete(ContractTypePo contractTypePo);
}
| 198 |
575 | <reponame>zealoussnow/chromium<gh_stars>100-1000
// Copyright 2018 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "chrome/elevation_service/elevated_recovery_impl.h"
#include <objbase.h>
#include <string>
#include <utility>
#include "base/base_paths.h"
#include "base/bind.h"
#include "base/callback_helpers.h"
#include "base/command_line.h"
#include "base/files/file_enumerator.h"
#include "base/files/file_path.h"
#include "base/files/file_util.h"
#include "base/files/scoped_temp_dir.h"
#include "base/logging.h"
#include "base/path_service.h"
#include "base/process/launch.h"
#include "base/process/process.h"
#include "base/strings/utf_string_conversions.h"
#include "base/version.h"
#include "base/win/scoped_process_information.h"
#include "chrome/install_static/install_util.h"
#include "third_party/zlib/google/zip.h"
namespace elevation_service {
namespace {
// Input CRX files over 10 MB are considered invalid.
constexpr int64_t kMaxFileSize = 10u * 1000u * 1000u;
// The hard-coded file name that the Recovery CRX is copied to.
constexpr base::FilePath::CharType kCRXFileName[] =
FILE_PATH_LITERAL("ChromeRecoveryCRX.crx");
// The hard-coded Recovery subdirectory where the CRX is unpacked and executed.
constexpr base::FilePath::CharType kRecoveryDirectory[] =
FILE_PATH_LITERAL("ChromeRecovery");
// The hard-coded Recovery executable. This file comes from the CRX, and we
// create an elevated process from it.
constexpr base::FilePath::CharType kRecoveryExeName[] =
FILE_PATH_LITERAL("ChromeRecovery.exe");
// The hard-coded SHA256 of the SubjectPublicKeyInfo used to sign the Recovery
// CRX which contains ChromeRecovery.exe.
std::vector<uint8_t> GetRecoveryCRXHash() {
return std::vector<uint8_t>{0x5f, 0x94, 0xe0, 0x3c, 0x64, 0x30, 0x9f, 0xbc,
0xfe, 0x00, 0x9a, 0x27, 0x3e, 0x52, 0xbf, 0xa5,
0x84, 0xb9, 0xb3, 0x75, 0x07, 0x29, 0xde, 0xfa,
0x32, 0x76, 0xd9, 0x93, 0xb5, 0xa3, 0xce, 0x02};
}
// This function is only meant to be called when a Windows API function errors
// out, and the corresponding ::GetLastResult is expected to be set to an error.
// There could be cases where ::GetLastError() is not set correctly, this
// function returns E_FAIL in those cases.
HRESULT HRESULTFromLastError() {
const auto error_code = ::GetLastError();
return (error_code != ERROR_SUCCESS) ? HRESULT_FROM_WIN32(error_code)
: E_FAIL;
}
// Opens and returns the COM caller's |process| given the process id, or the
// current process if |proc_id| is 0.
HRESULT OpenCallingProcess(uint32_t proc_id, base::Process* process) {
DCHECK(proc_id);
DCHECK(process);
HRESULT hr = ::CoImpersonateClient();
if (FAILED(hr))
return hr;
base::ScopedClosureRunner revert_to_self(
base::BindOnce([]() { ::CoRevertToSelf(); }));
*process = base::Process::OpenWithAccess(proc_id, PROCESS_DUP_HANDLE);
return process->IsValid() ? S_OK : HRESULTFromLastError();
}
// Opens and returns a base::File instance for the |file_path|. We impersonate
// the COM caller when opening the base::File instance. This is to ensure that
// the COM caller has access to the file.
HRESULT OpenFileImpersonated(const base::FilePath& file_path,
int flags,
base::File* file) {
DCHECK(file);
HRESULT hr = ::CoImpersonateClient();
if (FAILED(hr))
return hr;
base::ScopedClosureRunner revert_to_self(
base::BindOnce([]() { ::CoRevertToSelf(); }));
file->Initialize(file_path, flags);
if (!file->IsValid())
return HRESULTFromLastError();
if (::GetFileType(file->GetPlatformFile()) != FILE_TYPE_DISK)
return E_INVALIDARG;
int64_t from_length = file->GetLength();
return from_length > 0 && from_length < kMaxFileSize ? S_OK : E_INVALIDARG;
}
// Opens |from| while impersonating the COM caller, and then copies the contents
// of |from| to |to|. |from| is opened impersonated to ensure that the COM
// caller has access to the file.
HRESULT CopyFileImpersonated(const base::FilePath from,
const base::FilePath& to) {
base::File from_file;
HRESULT hr =
OpenFileImpersonated(from,
base::File::FLAG_READ | base::File::FLAG_OPEN |
base::File::FLAG_SEQUENTIAL_SCAN,
&from_file);
if (FAILED(hr))
return hr;
base::File to_file;
to_file.Initialize(to, base::File::FLAG_WRITE |
base::File::FLAG_CREATE_ALWAYS |
base::File::FLAG_SEQUENTIAL_SCAN);
if (!to_file.IsValid())
return HRESULTFromLastError();
constexpr size_t kBufferSize = 0x10000;
std::vector<char> buffer(kBufferSize);
for (uint64_t total_bytes_read = 0;;) {
const int bytes_read =
from_file.ReadAtCurrentPos(buffer.data(), buffer.size());
if (bytes_read < 0)
return HRESULTFromLastError();
if (bytes_read == 0)
return S_OK;
total_bytes_read += bytes_read;
if (total_bytes_read > kMaxFileSize)
return E_INVALIDARG;
const int bytes_written = to_file.WriteAtCurrentPos(&buffer[0], bytes_read);
if (bytes_written < 0)
return HRESULTFromLastError();
if (bytes_written != bytes_read)
return E_UNEXPECTED;
}
NOTREACHED();
return S_OK;
}
// Validates the provided CRX using the |crx_hash|, and if validation succeeds,
// unpacks the CRX under |unpack_under_path|. Returns the unpacked CRX
// directory in |unpacked_crx_dir|.
HRESULT ValidateAndUnpackCRX(const base::FilePath& from_crx_path,
const crx_file::VerifierFormat& crx_format,
const std::vector<uint8_t>& crx_hash,
const base::FilePath& unpack_under_path,
base::ScopedTempDir* unpacked_crx_dir) {
DCHECK(unpacked_crx_dir);
base::ScopedTempDir to_dir;
if (!to_dir.CreateUniqueTempDirUnderPath(unpack_under_path))
return HRESULTFromLastError();
const base::FilePath to_crx_path = to_dir.GetPath().Append(kCRXFileName);
// Copy |from_crx_path| impersonated. This is to prevent us from copying files
// that may not be accessible to the calling COM user.
HRESULT hr = CopyFileImpersonated(from_crx_path, to_crx_path);
if (FAILED(hr))
return hr;
std::string public_key;
if (crx_file::Verify(to_crx_path, crx_format, {crx_hash}, {}, &public_key,
nullptr, /*compressed_verified_contents=*/nullptr) !=
crx_file::VerifierResult::OK_FULL) {
return CRYPT_E_NO_MATCH;
}
if (!zip::Unzip(to_crx_path, to_dir.GetPath()))
return E_UNEXPECTED;
LOG_IF(WARNING, !base::DeleteFile(to_crx_path));
LOG_IF(WARNING, !unpacked_crx_dir->Set(to_dir.Take()));
return S_OK;
}
// Runs the executable |path_and_name| with the provided |args|. The returned
// |proc_handle| is a process handle that is valid for the |calling_process|
// process.
HRESULT LaunchCmd(const base::CommandLine& command_line,
const base::Process& calling_process,
base::win::ScopedHandle* proc_handle) {
DCHECK(!command_line.GetCommandLineString().empty());
DCHECK(proc_handle);
base::LaunchOptions options = {};
options.feedback_cursor_off = true;
base::GetTempDir(&options.current_directory);
base::Process proc = base::LaunchProcess(command_line, options);
if (!proc.IsValid())
return HRESULTFromLastError();
HANDLE duplicate_proc_handle = nullptr;
constexpr DWORD desired_access =
PROCESS_QUERY_LIMITED_INFORMATION | SYNCHRONIZE;
bool res = ::DuplicateHandle(
::GetCurrentProcess(), // Current process.
proc.Handle(), // Process handle to duplicate.
calling_process.Handle(), // Process receiving the handle.
&duplicate_proc_handle, // Duplicated handle.
desired_access, // Access requested for the new handle.
FALSE, // Don't inherit the new handle.
0) != 0;
if (!res)
return HRESULTFromLastError();
proc_handle->Set(duplicate_proc_handle);
return S_OK;
}
HRESULT ValidateCRXArgs(const std::wstring& browser_appid,
const std::wstring& browser_version,
const std::wstring& session_id) {
if (!browser_appid.empty()) {
GUID guid = {};
HRESULT hr = ::IIDFromString(browser_appid.c_str(), &guid);
if (FAILED(hr))
return hr;
}
const base::Version version(base::WideToASCII(browser_version));
if (!version.IsValid())
return E_INVALIDARG;
GUID session_guid = {};
return ::IIDFromString(session_id.c_str(), &session_guid);
}
// Deletes all the files and subdirectories within |directory_path|. Errors are
// ignored.
void DeleteDirectoryFiles(const base::FilePath& directory_path) {
base::FileEnumerator file_enum(
directory_path, false,
base::FileEnumerator::FILES | base::FileEnumerator::DIRECTORIES);
for (base::FilePath current = file_enum.Next(); !current.empty();
current = file_enum.Next()) {
base::DeletePathRecursively(current);
}
}
// Schedules deletion after reboot of |dir_name| as well as all the files and
// subdirectories within |dir_name|. Errors are ignored.
void ScheduleDirectoryForDeletion(const base::FilePath& dir_name) {
// First schedule all the files within |dir_name| for deletion.
base::FileEnumerator file_enum(dir_name, false, base::FileEnumerator::FILES);
for (base::FilePath file = file_enum.Next(); !file.empty();
file = file_enum.Next()) {
base::DeleteFileAfterReboot(file);
}
// Then recurse to all the subdirectories.
base::FileEnumerator dir_enum(dir_name, false,
base::FileEnumerator::DIRECTORIES);
for (base::FilePath sub_dir = dir_enum.Next(); !sub_dir.empty();
sub_dir = dir_enum.Next()) {
ScheduleDirectoryForDeletion(sub_dir);
}
// Now schedule the empty directory itself.
base::DeleteFileAfterReboot(dir_name);
}
// Returns the ChromeRecovery directory under Google\Chrome. For machine
// installs, this directory is under %ProgramFiles%, which is writeable only by
// adminstrators. We use this secure directory to validate and unpack the CRX to
// prevent tampering.
HRESULT GetChromeRecoveryDirectory(base::FilePath* dir) {
base::FilePath recovery_dir;
if (!base::PathService::Get(base::DIR_EXE, &recovery_dir))
return HRESULTFromLastError();
recovery_dir = recovery_dir.DirName().DirName().Append(kRecoveryDirectory);
*dir = std::move(recovery_dir);
return S_OK;
}
} // namespace
HRESULT CleanupChromeRecoveryDirectory() {
base::FilePath recovery_dir;
HRESULT hr = GetChromeRecoveryDirectory(&recovery_dir);
if (FAILED(hr))
return hr;
DeleteDirectoryFiles(recovery_dir);
return S_OK;
}
HRESULT RunChromeRecoveryCRX(const base::FilePath& crx_path,
const std::wstring& browser_appid,
const std::wstring& browser_version,
const std::wstring& session_id,
uint32_t caller_proc_id,
base::win::ScopedHandle* proc_handle) {
if (crx_path.empty() || !caller_proc_id || !proc_handle)
return E_INVALIDARG;
HRESULT hr = ValidateCRXArgs(browser_appid, browser_version, session_id);
if (FAILED(hr))
return hr;
base::CommandLine args(base::CommandLine::NO_PROGRAM);
if (!browser_appid.empty())
args.AppendSwitchNative("appguid", browser_appid);
args.AppendSwitchNative("browser-version", browser_version);
args.AppendSwitchNative("sessionid", session_id);
args.AppendSwitch("system");
base::FilePath unpack_dir;
hr = GetChromeRecoveryDirectory(&unpack_dir);
if (FAILED(hr))
return hr;
return RunCRX(crx_path, args,
crx_file::VerifierFormat::CRX3_WITH_PUBLISHER_PROOF,
GetRecoveryCRXHash(), unpack_dir,
base::FilePath(kRecoveryExeName), caller_proc_id, proc_handle);
}
HRESULT RunCRX(const base::FilePath& crx_path,
const base::CommandLine& args,
const crx_file::VerifierFormat& crx_format,
const std::vector<uint8_t>& crx_hash,
const base::FilePath& unpack_under_path,
const base::FilePath& exe_filename,
uint32_t caller_proc_id,
base::win::ScopedHandle* proc_handle) {
DCHECK(!crx_path.empty());
DCHECK(!crx_hash.empty());
DCHECK(!unpack_under_path.empty());
DCHECK(!exe_filename.empty());
DCHECK(caller_proc_id);
DCHECK(proc_handle);
base::Process calling_process;
HRESULT hr = OpenCallingProcess(caller_proc_id, &calling_process);
if (FAILED(hr))
return hr;
base::ScopedTempDir unpacked_crx_dir;
hr = ValidateAndUnpackCRX(crx_path, crx_format, crx_hash, unpack_under_path,
&unpacked_crx_dir);
if (FAILED(hr))
return hr;
const base::FilePath path_and_name =
unpacked_crx_dir.GetPath().Append(exe_filename);
base::CommandLine command_line(path_and_name);
command_line.AppendArguments(args, false);
base::win::ScopedHandle scoped_proc_handle;
hr = LaunchCmd(command_line, calling_process, &scoped_proc_handle);
if (FAILED(hr))
return hr;
ScheduleDirectoryForDeletion(unpacked_crx_dir.Take());
*proc_handle = std::move(scoped_proc_handle);
return hr;
}
} // namespace elevation_service
| 5,586 |
357 | # Copyright 2019 The Blueqat Developers
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from collections import Counter
import pytest
from blueqat.utils import to_inttuple
@pytest.mark.parametrize('arg, expect', [
("01011", (0, 1, 0, 1, 1)),
({"00011": 2, "10100": 3}, {(0, 0, 0, 1, 1): 2, (1, 0, 1, 0, 0): 3}),
(Counter({"00011": 2, "10100": 3}), Counter({(0, 0, 0, 1, 1): 2, (1, 0, 1, 0, 0): 3}))
])
def test_to_inttuple(arg, expect):
assert to_inttuple(arg) == expect
| 336 |
1,179 | // SPDX-License-Identifier: BSD-2-Clause
/*
* Copyright (c) 2014, STMicroelectronics International N.V.
*/
#include <compiler.h>
#include <link.h>
#include <stdbool.h>
#include <stdlib.h>
#include <string.h>
#include <sys/queue.h>
#include <tee_api.h>
#include <tee_ta_api.h>
#include <tee_internal_api_extensions.h>
#include <user_ta_header.h>
#include <utee_syscalls.h>
#include <tee_arith_internal.h>
#include <malloc.h>
#include "tee_api_private.h"
struct ta_session {
uint32_t session_id;
void *session_ctx;
TAILQ_ENTRY(ta_session) link;
};
static TAILQ_HEAD(ta_sessions, ta_session) ta_sessions =
TAILQ_HEAD_INITIALIZER(ta_sessions);
static bool init_done;
/* From user_ta_header.c, built within TA */
extern uint8_t ta_heap[];
extern const size_t ta_heap_size;
extern struct ta_head ta_head;
uint32_t ta_param_types;
TEE_Param ta_params[TEE_NUM_PARAMS];
struct __elf_phdr_info __elf_phdr_info;
struct phdr_info {
struct dl_phdr_info info;
TAILQ_ENTRY(phdr_info) link;
};
static TAILQ_HEAD(phdr_info_head, phdr_info) __phdr_info_head =
TAILQ_HEAD_INITIALIZER(__phdr_info_head);
/*
* Keep track of how many modules have been initialized so that subsequent
* dlopen() calls will not run the same initializers again
*/
static size_t _num_mod_init;
static int _init_iterate_phdr_cb(struct dl_phdr_info *info,
size_t size __unused, void *data)
{
struct phdr_info *qe = NULL;
size_t *count = data;
qe = malloc(sizeof(*qe));
if (!qe) {
EMSG("init/fini: out of memory");
abort();
}
qe->info = *info;
TAILQ_INSERT_TAIL(&__phdr_info_head, qe, link);
(*count)++;
return 0;
}
static void _get_fn_array(struct dl_phdr_info *info, Elf_Sword tag_a,
Elf_Sword tag_s, void (***fn)(void), size_t *num_fn)
{
const Elf_Phdr *phdr = NULL;
Elf_Dyn *dyn = NULL;
size_t num_dyn = 0;
size_t i = 0;
size_t j = 0;
for (i = 0; i < info->dlpi_phnum; i++) {
phdr = info->dlpi_phdr + i;
if (phdr->p_type != PT_DYNAMIC)
continue;
num_dyn = phdr->p_memsz / sizeof(Elf_Dyn);
dyn = (Elf_Dyn *)(phdr->p_vaddr + info->dlpi_addr);
for (j = 0; j < num_dyn; j++) {
if (*fn && *num_fn)
break;
if (dyn->d_tag == DT_NULL) {
break;
} else if (dyn->d_tag == tag_a) {
*fn = (void (**)(void))(dyn->d_un.d_ptr +
info->dlpi_addr);
} else if (dyn->d_tag == tag_s) {
*num_fn = dyn->d_un.d_val / sizeof(Elf_Addr);
}
dyn++;
}
}
}
void __utee_call_elf_init_fn(void)
{
void (**fn)(void) = NULL;
size_t num_mod = 0;
size_t num_fn = 0;
size_t mod = 0;
size_t i = 0;
struct phdr_info *qe = NULL;
struct phdr_info *qe2 = NULL;
dl_iterate_phdr(_init_iterate_phdr_cb, &num_mod);
/* Reverse order: dependencies first */
TAILQ_FOREACH_REVERSE(qe, &__phdr_info_head, phdr_info_head, link) {
if (mod == num_mod - _num_mod_init)
break;
_get_fn_array(&qe->info, DT_INIT_ARRAY, DT_INIT_ARRAYSZ, &fn,
&num_fn);
for (i = 0; i < num_fn; i++)
fn[i]();
fn = NULL;
num_fn = 0;
mod++;
}
_num_mod_init += mod;
TAILQ_FOREACH_SAFE(qe, &__phdr_info_head, link, qe2) {
TAILQ_REMOVE(&__phdr_info_head, qe, link);
free(qe);
}
}
static int _fini_iterate_phdr_cb(struct dl_phdr_info *info,
size_t size __unused, void *data __unused)
{
void (**fn)(void) = NULL;
size_t num_fn = 0;
size_t i = 0;
_get_fn_array(info, DT_FINI_ARRAY, DT_FINI_ARRAYSZ, &fn, &num_fn);
for (i = 1; i <= num_fn; i++)
fn[num_fn - i]();
return 0;
}
void __utee_call_elf_fini_fn(void)
{
dl_iterate_phdr(_fini_iterate_phdr_cb, NULL);
}
static TEE_Result init_instance(void)
{
trace_set_level(tahead_get_trace_level());
__utee_gprof_init();
malloc_add_pool(ta_heap, ta_heap_size);
_TEE_MathAPI_Init();
__utee_tcb_init();
__utee_call_elf_init_fn();
return TA_CreateEntryPoint();
}
static void uninit_instance(void)
{
__utee_gprof_fini();
TA_DestroyEntryPoint();
__utee_call_elf_fini_fn();
}
static void ta_header_save_params(uint32_t param_types,
TEE_Param params[TEE_NUM_PARAMS])
{
ta_param_types = param_types;
if (params)
memcpy(ta_params, params, sizeof(ta_params));
else
memset(ta_params, 0, sizeof(ta_params));
}
static struct ta_session *ta_header_get_session(uint32_t session_id)
{
struct ta_session *itr;
TAILQ_FOREACH(itr, &ta_sessions, link) {
if (itr->session_id == session_id)
return itr;
}
return NULL;
}
static TEE_Result ta_header_add_session(uint32_t session_id)
{
struct ta_session *itr = ta_header_get_session(session_id);
TEE_Result res;
if (itr)
return TEE_SUCCESS;
if (!init_done) {
init_done = true;
res = init_instance();
if (res)
return res;
}
itr = TEE_Malloc(sizeof(struct ta_session),
TEE_USER_MEM_HINT_NO_FILL_ZERO);
if (!itr)
return TEE_ERROR_OUT_OF_MEMORY;
itr->session_id = session_id;
itr->session_ctx = 0;
TAILQ_INSERT_TAIL(&ta_sessions, itr, link);
return TEE_SUCCESS;
}
static void ta_header_remove_session(uint32_t session_id)
{
struct ta_session *itr;
bool keep_alive;
TAILQ_FOREACH(itr, &ta_sessions, link) {
if (itr->session_id == session_id) {
TAILQ_REMOVE(&ta_sessions, itr, link);
TEE_Free(itr);
keep_alive =
(ta_head.flags & TA_FLAG_SINGLE_INSTANCE) &&
(ta_head.flags & TA_FLAG_INSTANCE_KEEP_ALIVE);
if (TAILQ_EMPTY(&ta_sessions) && !keep_alive)
uninit_instance();
return;
}
}
}
static void to_utee_params(struct utee_params *up, uint32_t param_types,
const TEE_Param params[TEE_NUM_PARAMS])
{
size_t n = 0;
up->types = param_types;
for (n = 0; n < TEE_NUM_PARAMS; n++) {
switch (TEE_PARAM_TYPE_GET(param_types, n)) {
case TEE_PARAM_TYPE_VALUE_INPUT:
case TEE_PARAM_TYPE_VALUE_OUTPUT:
case TEE_PARAM_TYPE_VALUE_INOUT:
up->vals[n * 2] = params[n].value.a;
up->vals[n * 2 + 1] = params[n].value.b;
break;
case TEE_PARAM_TYPE_MEMREF_INPUT:
case TEE_PARAM_TYPE_MEMREF_OUTPUT:
case TEE_PARAM_TYPE_MEMREF_INOUT:
up->vals[n * 2] = (uintptr_t)params[n].memref.buffer;
up->vals[n * 2 + 1] = params[n].memref.size;
break;
default:
up->vals[n * 2] = 0;
up->vals[n * 2 + 1] = 0;
break;
}
}
}
static void from_utee_params(TEE_Param params[TEE_NUM_PARAMS],
uint32_t *param_types,
const struct utee_params *up)
{
size_t n;
uint32_t types = up->types;
for (n = 0; n < TEE_NUM_PARAMS; n++) {
uintptr_t a = up->vals[n * 2];
uintptr_t b = up->vals[n * 2 + 1];
switch (TEE_PARAM_TYPE_GET(types, n)) {
case TEE_PARAM_TYPE_VALUE_INPUT:
case TEE_PARAM_TYPE_VALUE_OUTPUT:
case TEE_PARAM_TYPE_VALUE_INOUT:
params[n].value.a = a;
params[n].value.b = b;
break;
case TEE_PARAM_TYPE_MEMREF_INPUT:
case TEE_PARAM_TYPE_MEMREF_OUTPUT:
case TEE_PARAM_TYPE_MEMREF_INOUT:
params[n].memref.buffer = (void *)a;
params[n].memref.size = b;
break;
default:
break;
}
}
if (param_types)
*param_types = types;
}
static TEE_Result entry_open_session(unsigned long session_id,
struct utee_params *up)
{
TEE_Result res;
struct ta_session *session;
uint32_t param_types;
TEE_Param params[TEE_NUM_PARAMS];
res = ta_header_add_session(session_id);
if (res != TEE_SUCCESS)
return res;
session = ta_header_get_session(session_id);
if (!session)
return TEE_ERROR_BAD_STATE;
from_utee_params(params, ¶m_types, up);
ta_header_save_params(param_types, params);
res = TA_OpenSessionEntryPoint(param_types, params,
&session->session_ctx);
to_utee_params(up, param_types, params);
if (res != TEE_SUCCESS)
ta_header_remove_session(session_id);
return res;
}
static TEE_Result entry_close_session(unsigned long session_id)
{
struct ta_session *session = ta_header_get_session(session_id);
if (!session)
return TEE_ERROR_BAD_STATE;
TA_CloseSessionEntryPoint(session->session_ctx);
ta_header_remove_session(session_id);
return TEE_SUCCESS;
}
static TEE_Result entry_invoke_command(unsigned long session_id,
struct utee_params *up, unsigned long cmd_id)
{
TEE_Result res;
uint32_t param_types;
TEE_Param params[TEE_NUM_PARAMS];
struct ta_session *session = ta_header_get_session(session_id);
if (!session)
return TEE_ERROR_BAD_STATE;
from_utee_params(params, ¶m_types, up);
ta_header_save_params(param_types, params);
res = TA_InvokeCommandEntryPoint(session->session_ctx, cmd_id,
param_types, params);
to_utee_params(up, param_types, params);
return res;
}
TEE_Result __utee_entry(unsigned long func, unsigned long session_id,
struct utee_params *up, unsigned long cmd_id)
{
TEE_Result res;
switch (func) {
case UTEE_ENTRY_FUNC_OPEN_SESSION:
res = entry_open_session(session_id, up);
break;
case UTEE_ENTRY_FUNC_CLOSE_SESSION:
res = entry_close_session(session_id);
break;
case UTEE_ENTRY_FUNC_INVOKE_COMMAND:
res = entry_invoke_command(session_id, up, cmd_id);
break;
default:
res = 0xffffffff;
TEE_Panic(0);
break;
}
ta_header_save_params(0, NULL);
return res;
}
| 4,063 |
381 | <reponame>joyrun/ActivityRouter
package com.grouter;
import java.lang.annotation.ElementType;
import java.lang.annotation.Retention;
import java.lang.annotation.RetentionPolicy;
import java.lang.annotation.Target;
/**
* 默认使用空构造方法实现,如果带有参数的构造方法需要实现就要使用ComponentConstructor注解
* Created by Wiki on 19/8/14.
*/
@Target({ElementType.TYPE})
@Retention(RetentionPolicy.RUNTIME)
public @interface RouterComponent {
/**
* path
*/
String value() default "";
/**
* 接口
*/
Class protocol();
/**
* 用于文档,名称
*/
String name() default "";
/**
* 用于文档,描述
*/
String description() default "";
}
| 341 |
190,993 | /* Copyright 2017 The TensorFlow Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
==============================================================================*/
#include <cmath>
#include "tensorflow/core/platform/logging.h"
#include "tensorflow/lite/toco/graph_transformations/graph_transformations.h"
#include "tensorflow/lite/toco/model.h"
#include "tensorflow/lite/toco/tooling_util.h"
namespace toco {
template <ArrayDataType A, typename T>
void FillRangeOutput(const Array& start_array, const Array& limit_array,
const Array& delta_array, Array* output_array) {
// Compute buffer contents
T start = start_array.GetBuffer<A>().data[0];
T limit = limit_array.GetBuffer<A>().data[0];
T delta = delta_array.GetBuffer<A>().data[0];
auto& buffer = output_array->GetMutableBuffer<A>();
buffer.data.clear();
int size =
(std::is_integral<T>::value
? ((std::abs(limit - start) + std::abs(delta) - 1) / std::abs(delta))
: std::ceil(std::abs((limit - start) / delta)));
for (int i = 0; i < size; ++i) {
buffer.data.push_back(start + i * delta);
}
CHECK_EQ(std::floor((limit - start) / delta), buffer.data.size());
CHECK_EQ(buffer.data.size(), output_array->shape().dims()[0]);
}
::tensorflow::Status ResolveConstantRange::Run(Model* model,
std::size_t op_index,
bool* modified) {
*modified = false;
const auto it = model->operators.begin() + op_index;
auto* base_op = it->get();
if (base_op->type != OperatorType::kRange) {
return ::tensorflow::Status::OK();
}
auto* op = static_cast<RangeOperator*>(base_op);
CHECK_EQ(op->inputs.size(), 3);
const auto& start_array = model->GetArray(op->inputs[0]);
if (!start_array.has_shape()) {
// Yield until all input dims have been resolved.
return ::tensorflow::Status::OK();
}
const auto& limit_array = model->GetArray(op->inputs[1]);
if (!limit_array.has_shape()) {
// Yield until all input dims have been resolved.
return ::tensorflow::Status::OK();
}
const auto& delta_array = model->GetArray(op->inputs[2]);
if (!delta_array.has_shape()) {
// Yield until all input dims have been resolved.
return ::tensorflow::Status::OK();
}
for (const auto& input : op->inputs) {
if (!IsConstantParameterArray(*model, input)) {
// yield if any input is mutable
return ::tensorflow::Status::OK();
}
}
CHECK_EQ(op->outputs.size(), 1);
auto& output_array = model->GetArray(op->outputs[0]);
if (output_array.data_type == ArrayDataType::kNone) {
// Yield until the output type has been set by PropagateArrayDataTypes
return ::tensorflow::Status::OK();
}
CHECK_EQ(RequiredBufferSizeForShape(start_array.shape()), 1)
<< "Range op inputs must be scalar.";
CHECK_EQ(RequiredBufferSizeForShape(limit_array.shape()), 1)
<< "Range op inputs must be scalar.";
CHECK_EQ(RequiredBufferSizeForShape(delta_array.shape()), 1)
<< "Range op inputs must be scalar.";
CHECK(start_array.data_type == ArrayDataType::kInt32 ||
start_array.data_type == ArrayDataType::kFloat)
<< "Range op inputs must be int32 or float.";
CHECK(limit_array.data_type == start_array.data_type)
<< "Range op inputs type must be equal.";
CHECK(delta_array.data_type == start_array.data_type)
<< "Range op inputs type must be equal.";
if (start_array.data_type == ArrayDataType::kInt32) {
FillRangeOutput<ArrayDataType::kInt32, int32_t>(start_array, limit_array,
delta_array, &output_array);
} else {
FillRangeOutput<ArrayDataType::kFloat, float>(start_array, limit_array,
delta_array, &output_array);
}
DeleteOpAndArrays(model, op);
*modified = true;
return ::tensorflow::Status::OK();
}
} // namespace toco
| 1,692 |
1,679 | <filename>src/boards/mcu/saml21/hal/include/hpl_gpio.h
/**
* \file
*
* \brief Port related functionality declaration.
*
* Copyright (C) 2014-2016 Atmel Corporation. All rights reserved.
*
* \asf_license_start
*
* \page License
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
*
* 1. Redistributions of source code must retain the above copyright notice,
* this list of conditions and the following disclaimer.
*
* 2. Redistributions in binary form must reproduce the above copyright notice,
* this list of conditions and the following disclaimer in the documentation
* and/or other materials provided with the distribution.
*
* 3. The name of Atmel may not be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* 4. This software may only be redistributed and used in connection with an
* Atmel microcontroller product.
*
* THIS SOFTWARE IS PROVIDED BY ATMEL "AS IS" AND ANY EXPRESS OR IMPLIED
* WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT ARE
* EXPRESSLY AND SPECIFICALLY DISCLAIMED. IN NO EVENT SHALL ATMEL BE LIABLE FOR
* ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
* STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
* ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*
* \asf_license_stop
*
*/
#ifndef _HPL_GPIO_H_INCLUDED
#define _HPL_GPIO_H_INCLUDED
/**
* \addtogroup HPL Port
*
* \section hpl_port_rev Revision History
* - v1.0.0 Initial Release
*
*@{
*/
#include <compiler.h>
#ifdef __cplusplus
extern "C" {
#endif
/**
* \brief Macros for the pin and port group, lower 5
* bits stands for pin number in the group, higher 3
* bits stands for port group
*/
#define GPIO_PIN(n) (((n)&0x1Fu) << 0)
#define GPIO_PORT(n) ((n) >> 5)
#define GPIO(port, pin) ((((port)&0x7u) << 5) + ((pin)&0x1Fu))
#define GPIO_PIN_FUNCTION_OFF 0xffffffff
/**
* \brief PORT pull mode settings
*/
enum gpio_pull_mode { GPIO_PULL_OFF, GPIO_PULL_UP, GPIO_PULL_DOWN };
/**
* \brief PORT direction settins
*/
enum gpio_direction { GPIO_DIRECTION_OFF, GPIO_DIRECTION_IN, GPIO_DIRECTION_OUT };
/**
* \brief PORT group abstraction
*/
enum gpio_port { GPIO_PORTA, GPIO_PORTB, GPIO_PORTC, GPIO_PORTD, GPIO_PORTE };
/**
* \name HPL functions
*/
//@{
/**
* \brief Port initialization function
*
* Port initialization function should setup the port module based
* on a static configuration file, this function should normally
* not be called directly, but is a part of hal_init()
*/
void _gpio_init(void);
/**
* \brief Set direction on port with mask
*
* Set data direction for each pin, or disable the pin
*
* \param[in] port Ports are grouped into groups of maximum 32 pins,
* GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc
* \param[in] mask Bit mask where 1 means apply direction setting to the
* corresponding pin
* \param[in] direction GPIO_DIRECTION_OFF = set pin direction to input
* and disable input buffer to disable the pin
* GPIO_DIRECTION_IN = set pin direction to input
* and enable input buffer to enable the pin
* GPIO_DIRECTION_OUT = set pin direction to output
* and disable input buffer
*/
static inline void _gpio_set_direction(const enum gpio_port port, const uint32_t mask,
const enum gpio_direction direction);
/**
* \brief Set output level on port with mask
*
* Sets output state on pin to high or low with pin masking
*
* \param[in] port Ports are grouped into groups of maximum 32 pins,
* GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc
* \param[in] mask Bit mask where 1 means apply direction setting to
* the corresponding pin
* \param[in] level true = pin level is set to 1
* false = pin level is set to 0
*/
static inline void _gpio_set_level(const enum gpio_port port, const uint32_t mask, const bool level);
/**
* \brief Change output level to the opposite with mask
*
* Change pin output level to the opposite with pin masking
*
* \param[in] port Ports are grouped into groups of maximum 32 pins,
* GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc
* \param[in] mask Bit mask where 1 means apply direction setting to
* the corresponding pin
*/
static inline void _gpio_toggle_level(const enum gpio_port port, const uint32_t mask);
/**
* \brief Get input levels on all port pins
*
* Get input level on all port pins, will read IN register if configured to
* input and OUT register if configured as output
*
* \param[in] port Ports are grouped into groups of maximum 32 pins,
* GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc
*/
static inline uint32_t _gpio_get_level(const enum gpio_port port);
/**
* \brief Set pin pull mode
*
* Set pull mode on a single pin
*
* \notice This function will automatically change pin direction to input
*
* \param[in] port Ports are grouped into groups of maximum 32 pins,
* GPIO_PORTA = group 0, GPIO_PORTB = group 1, etc
* \param[in] pin The pin in the group that pull mode should be selected
* for
* \param[in] pull_mode GPIO_PULL_OFF = pull resistor on pin is disabled
* GPIO_PULL_DOWN = pull resistor on pin will pull pin
* level to ground level
* GPIO_PULL_UP = pull resistor on pin will pull pin
* level to VCC
*/
static inline void _gpio_set_pin_pull_mode(const enum gpio_port port, const uint8_t pin,
const enum gpio_pull_mode pull_mode);
/**
* \brief Set gpio function
*
* Select which function a gpio is used for
*
* \param[in] gpio The gpio to set function for
* \param[in] function The gpio function is given by a 32-bit wide bitfield
* found in the header files for the device
*
*/
static inline void _gpio_set_pin_function(const uint32_t gpio, const uint32_t function);
#include <hpl_gpio_base.h>
//@}
#ifdef __cplusplus
}
#endif
/**@}*/
#endif /* _HPL_GPIO_H_INCLUDED */
| 2,449 |
487 | print( f"{dumps(data)}\n" ) | 14 |
1,355 | <gh_stars>1000+
// Copyright (c) 2018 <NAME>
//
// I am making my contributions/submissions to this project solely in my
// personal capacity and am not conveying any rights to any intellectual
// property of any third parties.
#include "points_to_implicit.h"
#include "pybind11_utils.h"
#include <jet/points_to_implicit2.h>
#include <jet/points_to_implicit3.h>
namespace py = pybind11;
using namespace jet;
void addPointsToImplicit2(pybind11::module& m) {
py::class_<PointsToImplicit2, PointsToImplicit2Ptr>(m, "PointsToImplicit2",
R"pbdoc(
Abstract base class for 2-D points-to-implicit converters.
)pbdoc")
.def("convert",
[](const PointsToImplicit2& instance, const py::list& points,
ScalarGrid2Ptr output) {
std::vector<Vector2D> points_;
for (size_t i = 0; i < points.size(); ++i) {
points_.push_back(objectToVector2D(points[i]));
}
ConstArrayAccessor1<Vector2D> pointsAcc(points_.size(),
points_.data());
instance.convert(pointsAcc, output.get());
},
R"pbdoc(
Converts the given points to implicit surface scalar field.
Parameters
----------
- points : List of 2D vectors.
- output : Scalar grid output.
)pbdoc",
py::arg("points"), py::arg("output"));
}
void addPointsToImplicit3(pybind11::module& m) {
py::class_<PointsToImplicit3, PointsToImplicit3Ptr>(m, "PointsToImplicit3",
R"pbdoc(
Abstract base class for 3-D points-to-implicit converters.
)pbdoc")
.def("convert",
[](const PointsToImplicit3& instance, const py::list& points,
ScalarGrid3Ptr output) {
std::vector<Vector3D> points_;
for (size_t i = 0; i < points.size(); ++i) {
points_.push_back(objectToVector3D(points[i]));
}
ConstArrayAccessor1<Vector3D> pointsAcc(points_.size(),
points_.data());
instance.convert(pointsAcc, output.get());
},
R"pbdoc(
Converts the given points to implicit surface scalar field.
Parameters
----------
- points : List of 3D vectors.
- output : Scalar grid output.
)pbdoc",
py::arg("points"), py::arg("output"));
}
| 1,400 |
1,279 | /*
* Copyright (c) 2012-2021 <NAME> et al.
* License: https://github.com/dbartolini/crown/blob/master/LICENSE
*/
#pragma once
#include "core/filesystem/types.h"
#include "core/memory/types.h"
#include "core/types.h"
#include "resource/types.h"
namespace crown
{
namespace config_resource_internal
{
s32 compile(CompileOptions& opts);
void* load(File& file, Allocator& a);
void unload(Allocator& allocator, void* resource);
} // namespace config_resource_internal
} // namespace crown
| 177 |
558 | # Copyright 2018 The TensorFlow Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may not
# use this file except in compliance with the License. You may obtain a copy of
# the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations under
# the License.
# ==============================================================================
"""test ffmpeg dataset"""
import os
import sys
import pytest
import tensorflow as tf
import tensorflow_io as tfio
@pytest.fixture(name="video_data", scope="module")
def fixture_video_data():
"""fixture_video_data"""
path = "file://" + os.path.join(
os.path.dirname(os.path.abspath(__file__)), "test_video", "small.mp4"
)
# TODO: get raw value
value = tf.zeros((166, 320, 560, 3), tf.uint8)
return path, value
@pytest.mark.parametrize(
("io_dataset_func"),
[
pytest.param(lambda f: tfio.IODataset.graph(tf.uint8).from_ffmpeg(f, "v:0")),
pytest.param(lambda f: tfio.IODataset.from_ffmpeg(f, "v:0")),
],
ids=["from_ffmpeg", "from_ffmpeg(eager)"],
)
def test_video_io_dataset(video_data, io_dataset_func):
"""test_video_io_dataset"""
video_path, video_value = video_data
video_dataset = io_dataset_func(video_path)
i = 0
for value in video_dataset:
assert video_value[i].shape == value.shape
i += 1
assert i == 166
video_dataset = io_dataset_func(video_path).batch(2)
i = 0
for value in video_dataset:
assert video_value[i : i + 2].shape == value.shape
i += 2
assert i == 166
@pytest.mark.parametrize(
("io_dataset_func"),
[pytest.param(lambda f: tfio.IODataset.graph(tf.uint8).from_ffmpeg(f, "v:0"),)],
ids=["from_ffmpeg"],
)
def test_video_io_dataset_with_dataset(video_data, io_dataset_func):
"""test_video_io_dataset_with_dataset"""
video_path, video_value = video_data
filename_dataset = tf.data.Dataset.from_tensor_slices([video_path, video_path])
position_dataset = tf.data.Dataset.from_tensor_slices(
[tf.constant(50, tf.int64), tf.constant(100, tf.int64)]
)
dataset = tf.data.Dataset.zip((filename_dataset, position_dataset))
# Note: @tf.function is actually not needed, as tf.data.Dataset
# will automatically wrap the `func` into a graph anyway.
# The following is purely for explanation purposes.
# Return: an embedded dataset (in an outer dataset) for position:position+100
@tf.function
def func(filename, position):
video_dataset = io_dataset_func(filename)
return video_dataset.skip(position).take(10)
dataset = dataset.map(func)
item = 0
# Notice video_dataset in dataset:
for video_dataset in dataset:
position = 50 if item == 0 else 100
i = 0
for value in video_dataset:
assert video_value[position + i].shape == value.shape
i += 1
assert i == 10
item += 1
assert item == 2
| 1,292 |
310 | <filename>gear/hardware/0/600p-crl.json
{
"name": "600P CRL",
"description": "A glass cutter.",
"url": "https://www.amazon.com/LAURENCE-Heavy-Duty-Oil-Type-Pattern-Cutting/dp/B006JFN2OG"
} | 83 |
309 | // Copyright 2019 the deepx authors.
// Author: <NAME> (<EMAIL>)
//
#include <deepx_core/graph/op_impl.h>
namespace deepx_core {
namespace {
bool SequenceMaskInferShape(const Shape& X, int max_size, Shape* Z) noexcept {
if (!X.is_rank(1)) {
DXERROR("Invalid X: rank of X %d must be 1.", X.rank());
return false;
}
if (max_size <= 0) {
DXERROR("Invalid max_size: max_size %d must be positive.", max_size);
return false;
}
Z->resize(X[0], max_size);
return true;
}
template <typename T>
void SequenceMask(const Tensor<T>& X, Tensor<T>* Z) noexcept {
const T* _X = X.data();
T* _Z = Z->data();
int batch = Z->dim(0);
int m = Z->dim(1);
int Xi;
Z->zeros();
for (int i = 0; i < batch; ++i) {
Xi = (int)*_X;
if (Xi > 0) {
if (Xi > m) {
Xi = m;
}
for (int j = 0; j < Xi; ++j) {
_Z[j] = 1;
}
}
_X += 1;
_Z += m;
}
}
} // namespace
SequenceMaskNode::SequenceMaskNode(std::string name, GraphNode* X, int max_size)
: GraphNodeUnaryBase(std::move(name), X), max_size_(max_size) {
if (!X->shape().empty()) {
(void)SequenceMaskInferShape(X->shape(), max_size_, &shape_);
}
}
class SequenceMaskOp : public OpUnaryBase {
private:
Shape Zshape_;
public:
DEFINE_OP_LIKE(SequenceMaskOp);
const Shape& InferShape() override {
int max_size = ((const SequenceMaskNode*)node_)->max_size();
DXCHECK_THROW(SequenceMaskInferShape(X_->shape(), max_size, &Zshape_));
return Zshape_;
}
void Forward() override { SequenceMask(*X_, Z_); }
};
GRAPH_NODE_OP_REGISTER(SequenceMask);
} // namespace deepx_core
| 697 |
665 | <filename>3D_Printed_LED_Microphone_Flag/code.py<gh_stars>100-1000
# SPDX-FileCopyrightText: 2013 <NAME> for Adafruit Industries
# SPDX-FileCopyrightText: 2017 Mi<NAME> for Adafruit Industries
#
# SPDX-License-Identifier: BSD
# LED VU meter for Arduino and Adafruit NeoPixel LEDs.
# Hardware requirements:
# - M0 boards
# - Adafruit Electret Microphone Amplifier (ID: 1063)
# - Adafruit Flora RGB Smart Pixels (ID: 1260)
# OR
# - Adafruit NeoPixel Digital LED strip (ID: 1138)
# - Optional: battery for portable use (else power through USB or adapter)
# Software requirements:
# - Adafruit NeoPixel library
# Connections:
# - 3.3V to mic amp +
# - GND to mic amp -
# - Analog pin to microphone output (configurable below)
# - Digital pin to LED data input (configurable below)
# See notes in setup() regarding 5V vs. 3.3V boards - there may be an
# extra connection to make and one line of code to enable or disable.
# Written by Adafruit Industries. Distributed under the BSD license.
# This paragraph must be included in any redistribution.
# fscale function:
# Floating Point Autoscale Function V0.1
# Written by <NAME> 2007
# Modified fromhere code by <NAME>
# Ported to Circuit Python by <NAME>
import time
import board
import neopixel
from rainbowio import colorwheel
from analogio import AnalogIn
n_pixels = 16 # Number of pixels you are using
mic_pin = AnalogIn(board.A1) # Microphone is attached to this analog pin
led_pin = board.D1 # NeoPixel LED strand is connected to this pin
sample_window = .1 # Sample window for average level
peak_hang = 24 # Time of pause before peak dot falls
peak_fall = 4 # Rate of falling peak dot
input_floor = 10 # Lower range of analogRead input
# Max range of analogRead input, the lower the value the more sensitive
# (1023 = max)
input_ceiling = 300
peak = 16 # Peak level of column; used for falling dots
sample = 0
dotcount = 0 # Frame counter for peak dot
dothangcount = 0 # Frame counter for holding peak dot
strip = neopixel.NeoPixel(led_pin, n_pixels, brightness=1, auto_write=False)
def remapRange(value, leftMin, leftMax, rightMin, rightMax):
# this remaps a value fromhere original (left) range to new (right) range
# Figure out how 'wide' each range is
leftSpan = leftMax - leftMin
rightSpan = rightMax - rightMin
# Convert the left range into a 0-1 range (int)
valueScaled = int(value - leftMin) / int(leftSpan)
# Convert the 0-1 range into a value in the right range.
return int(rightMin + (valueScaled * rightSpan))
def fscale(originalmin, originalmax, newbegin, newend, inputvalue, curve):
invflag = 0
# condition curve parameter
# limit range
if curve > 10:
curve = 10
if curve < -10:
curve = -10
# - invert and scale -
# this seems more intuitive
# postive numbers give more weight to high end on output
curve = (curve * -.1)
# convert linear scale into lograthimic exponent for other pow function
curve = pow(10, curve)
# Check for out of range inputValues
if inputvalue < originalmin:
inputvalue = originalmin
if inputvalue > originalmax:
inputvalue = originalmax
# Zero Refference the values
originalrange = originalmax - originalmin
if newend > newbegin:
newrange = newend - newbegin
else:
newrange = newbegin - newend
invflag = 1
zerorefcurval = inputvalue - originalmin
# normalize to 0 - 1 float
normalizedcurval = zerorefcurval / originalrange
# Check for originalMin > originalMax
# -the math for all other cases
# i.e. negative numbers seems to work out fine
if originalmin > originalmax:
return 0
if invflag == 0:
rangedvalue = (pow(normalizedcurval, curve) * newrange) + newbegin
else: # invert the ranges
rangedvalue = newbegin - (pow(normalizedcurval, curve) * newrange)
return rangedvalue
def drawLine(fromhere, to):
if fromhere > to:
to, fromhere = fromhere, to
for index in range(fromhere, to):
strip[index] = (0, 0, 0)
while True:
time_start = time.monotonic() # current time used for sample window
peaktopeak = 0 # peak-to-peak level
signalmax = 0
signalmin = 1023
c = 0
y = 0
# collect data for length of sample window (in seconds)
while (time.monotonic() - time_start) < sample_window:
# convert to arduino 10-bit [1024] fromhere 16-bit [65536]
sample = mic_pin.value / 64
if sample < 1024: # toss out spurious readings
if sample > signalmax:
signalmax = sample # save just the max levels
elif sample < signalmin:
signalmin = sample # save just the min levels
peaktopeak = signalmax - signalmin # max - min = peak-peak amplitude
# Fill the strip with rainbow gradient
for i in range(0, len(strip)):
strip[i] = colorwheel(remapRange(i, 0, (n_pixels - 1), 30, 150))
# Scale the input logarithmically instead of linearly
c = fscale(input_floor, input_ceiling, (n_pixels - 1), 0, peaktopeak, 2)
if c < peak:
peak = c # keep dot on top
dothangcount = 0 # make the dot hang before falling
if c <= n_pixels: # fill partial column with off pixels
drawLine(n_pixels, n_pixels - int(c))
# Set the peak dot to match the rainbow gradient
y = n_pixels - peak
strip.fill = (y - 1, colorwheel(remapRange(y, 0, (n_pixels - 1), 30, 150)))
strip.write()
# Frame based peak dot animation
if dothangcount > peak_hang: # Peak pause length
dotcount += 1
if dotcount >= peak_fall: # Fall rate
peak += 1
dotcount = 0
else:
dothangcount += 1
| 2,039 |
348 | <filename>docs/data/t2/011/11335.json<gh_stars>100-1000
{"nom":"Sainte-Colombe-sur-Guette","dpt":"Aude","inscrits":50,"abs":6,"votants":44,"blancs":10,"nuls":1,"exp":33,"res":[{"panneau":"1","voix":25},{"panneau":"2","voix":8}]} | 101 |
891 | from typing import ClassVar
from datamodel_code_generator.model.pydantic.base_model import BaseModel
class CustomRootType(BaseModel):
TEMPLATE_FILE_PATH: ClassVar[str] = 'pydantic/BaseModel_root.jinja2'
BASE_CLASS: ClassVar[str] = 'pydantic.BaseModel'
| 96 |
2,293 | <gh_stars>1000+
/*****************************************************************************
* *
* This file is part of the BeanShell Java Scripting distribution. *
* Documentation and updates may be found at http://www.beanshell.org/ *
* *
* Sun Public License Notice: *
* *
* The contents of this file are subject to the Sun Public License Version *
* 1.0 (the "License"); you may not use this file except in compliance with *
* the License. A copy of the License is available at http://www.sun.com *
* *
* The Original Code is BeanShell. The Initial Developer of the Original *
* Code is <NAME>. Portions created by <NAME> are Copyright *
* (C) 2000. All Rights Reserved. *
* *
* GNU Public License Notice: *
* *
* Alternatively, the contents of this file may be used under the terms of *
* the GNU Lesser General Public License (the "LGPL"), in which case the *
* provisions of LGPL are applicable instead of those above. If you wish to *
* allow use of your version of this file only under the terms of the LGPL *
* and not to allow others to use your version of this file under the SPL, *
* indicate your decision by deleting the provisions above and replace *
* them with the notice and other provisions required by the LGPL. If you *
* do not delete the provisions above, a recipient may use your version of *
* this file under either the SPL or the LGPL. *
* *
* <NAME> (<EMAIL>) *
* Author of Learning Java, O'Reilly & Associates *
* http://www.pat.net/~pat/ *
* *
*****************************************************************************/
package bsh;
import java.io.*;
import java.net.*;
import java.text.*;
/**
Remote executor class. Posts a script from the command line to a BshServlet
or embedded interpreter using (respectively) HTTP or the bsh telnet
service. Output is printed to stdout and a numeric return value is scraped
from the result.
*/
public class Remote
{
public static void main( String args[] )
throws Exception
{
if ( args.length < 2 ) {
System.out.println(
"usage: Remote URL(http|bsh) file [ file ] ... ");
System.exit(1);
}
String url = args[0];
String text = getFile(args[1]);
int ret = eval( url, text );
System.exit( ret );
}
/**
Evaluate text in the interpreter at url, returning a possible integer
return value.
*/
public static int eval( String url, String text )
throws IOException
{
String returnValue = null;
if ( url.startsWith( "http:" ) ) {
returnValue = doHttp( url, text );
} else if ( url.startsWith( "bsh:" ) ) {
returnValue = doBsh( url, text );
} else
throw new IOException( "Unrecognized URL type."
+"Scheme must be http:// or bsh://");
try {
return Integer.parseInt( returnValue );
} catch ( Exception e ) {
// this convention may change...
return 0;
}
}
static String doBsh( String url, String text )
{
OutputStream out;
InputStream in;
String host = "";
String port = "";
String returnValue = "-1";
String orgURL = url;
// Need some format checking here
try {
url = url.substring(6); // remove the bsh://
// get the index of the : between the host and the port is located
int index = url.indexOf(":");
host = url.substring(0,index);
port = url.substring(index+1,url.length());
} catch ( Exception ex ) {
System.err.println("Bad URL: "+orgURL+": "+ex );
return returnValue;
}
try {
System.out.println("Connecting to host : "
+ host + " at port : " + port);
Socket s = new Socket(host, Integer.parseInt(port) + 1);
out = s.getOutputStream();
in = s.getInputStream();
sendLine( text, out );
BufferedReader bin = new BufferedReader(
new InputStreamReader(in));
String line;
while ( (line=bin.readLine()) != null )
System.out.println( line );
// Need to scrape a value from the last line?
returnValue="1";
return returnValue;
} catch(Exception ex) {
System.err.println("Error communicating with server: "+ex);
return returnValue;
}
}
private static void sendLine( String line, OutputStream outPipe )
throws IOException
{
outPipe.write( line.getBytes() );
outPipe.flush();
}
/*
TODO: this is not unicode friendly, nor is getFile()
The output is urlencoded 8859_1 text.
should probably be urlencoded UTF-8... how does the servlet determine
the encoded charset? I guess we're supposed to add a ";charset" clause
to the content type?
*/
static String doHttp( String postURL, String text )
{
String returnValue = null;
StringBuffer sb = new StringBuffer();
sb.append( "bsh.client=Remote" );
sb.append( "&bsh.script=" );
sb.append( URLEncoder.encode( text ) );
/*
// This requires Java 1.3
try {
sb.append( URLEncoder.encode( text, "8859_1" ) );
} catch ( UnsupportedEncodingException e ) {
e.printStackTrace();
}
*/
String formData = sb.toString( );
try {
URL url = new URL( postURL );
HttpURLConnection urlcon =
(HttpURLConnection) url.openConnection( );
urlcon.setRequestMethod("POST");
urlcon.setRequestProperty("Content-type",
"application/x-www-form-urlencoded");
urlcon.setDoOutput(true);
urlcon.setDoInput(true);
PrintWriter pout = new PrintWriter( new OutputStreamWriter(
urlcon.getOutputStream(), "8859_1"), true );
pout.print( formData );
pout.flush();
// read results...
int rc = urlcon.getResponseCode();
if ( rc != HttpURLConnection.HTTP_OK )
System.out.println("Error, HTTP response: "+rc );
returnValue = urlcon.getHeaderField("Bsh-Return");
BufferedReader bin = new BufferedReader(
new InputStreamReader( urlcon.getInputStream() ) );
String line;
while ( (line=bin.readLine()) != null )
System.out.println( line );
System.out.println( "Return Value: "+returnValue );
} catch (MalformedURLException e) {
System.out.println(e); // bad postURL
} catch (IOException e2) {
System.out.println(e2); // I/O error
}
return returnValue;
}
/*
Note: assumes default character encoding
*/
static String getFile( String name )
throws FileNotFoundException, IOException
{
StringBuffer sb = new StringBuffer();
BufferedReader bin = new BufferedReader( new FileReader( name ) );
String line;
while ( (line=bin.readLine()) != null )
sb.append( line ).append( "\n" );
return sb.toString();
}
}
| 3,130 |
653 |
//==---------------- opencl.hpp - SYCL OpenCL backend ----------------------==//
//
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
// See https://llvm.org/LICENSE.txt for license information.
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
//
//===----------------------------------------------------------------------===//
#pragma once
#include <CL/sycl/accessor.hpp>
#include <CL/sycl/backend.hpp>
#include <CL/sycl/backend_types.hpp>
#include <CL/sycl/detail/backend_traits.hpp>
#include <CL/sycl/detail/cl.h>
#include <CL/sycl/kernel_bundle.hpp>
#include <vector>
__SYCL_INLINE_NAMESPACE(cl) {
namespace sycl {
template <> struct interop<backend::opencl, platform> {
using type = cl_platform_id;
};
template <> struct interop<backend::opencl, device> {
using type = cl_device_id;
};
template <> struct interop<backend::opencl, context> {
using type = cl_context;
};
template <> struct interop<backend::opencl, queue> {
using type = cl_command_queue;
};
#ifdef __SYCL_INTERNAL_API
template <> struct interop<backend::opencl, program> {
using type = cl_program;
};
#endif
template <> struct interop<backend::opencl, event> { using type = cl_event; };
template <typename DataT, int Dimensions, access::mode AccessMode>
struct interop<backend::opencl,
accessor<DataT, Dimensions, AccessMode, access::target::device,
access::placeholder::false_t>> {
using type = cl_mem;
};
template <typename DataT, int Dimensions, access::mode AccessMode>
struct interop<backend::opencl, accessor<DataT, Dimensions, AccessMode,
access::target::constant_buffer,
access::placeholder::false_t>> {
using type = cl_mem;
};
template <typename DataT, int Dimensions, access::mode AccessMode>
struct interop<backend::opencl,
accessor<DataT, Dimensions, AccessMode, access::target::image,
access::placeholder::false_t>> {
using type = cl_mem;
};
template <typename DataT, int Dimensions, typename AllocatorT>
struct interop<backend::opencl, buffer<DataT, Dimensions, AllocatorT>> {
using type = cl_mem;
};
namespace detail {
template <bundle_state State>
struct BackendInput<backend::opencl, kernel_bundle<State>> {
using type = cl_program;
};
template <bundle_state State>
struct BackendReturn<backend::opencl, kernel_bundle<State>> {
using type = std::vector<cl_program>;
};
template <> struct BackendInput<backend::opencl, kernel> {
using type = cl_kernel;
};
template <> struct BackendReturn<backend::opencl, kernel> {
using type = cl_kernel;
};
template <> struct InteropFeatureSupportMap<backend::opencl> {
static constexpr bool MakePlatform = true;
static constexpr bool MakeDevice = true;
static constexpr bool MakeContext = true;
static constexpr bool MakeQueue = true;
static constexpr bool MakeEvent = true;
static constexpr bool MakeBuffer = true;
static constexpr bool MakeKernel = true;
static constexpr bool MakeKernelBundle = true;
};
} // namespace detail
namespace opencl {
// Implementation of various "make" functions resides in SYCL RT because
// creating SYCL objects requires knowing details not acessible here.
// Note that they take opaque pi_native_handle that real OpenCL handles
// are casted to.
//
__SYCL_EXPORT platform make_platform(pi_native_handle NativeHandle);
__SYCL_EXPORT device make_device(pi_native_handle NativeHandle);
__SYCL_EXPORT context make_context(pi_native_handle NativeHandle);
#ifdef __SYCL_INTERNAL_API
__SYCL_EXPORT program make_program(const context &Context,
pi_native_handle NativeHandle);
#endif
__SYCL_EXPORT queue make_queue(const context &Context,
pi_native_handle InteropHandle);
// Construction of SYCL platform.
template <typename T, typename detail::enable_if_t<
std::is_same<T, platform>::value> * = nullptr>
T make(typename interop<backend::opencl, T>::type Interop) {
return make_platform(detail::pi::cast<pi_native_handle>(Interop));
}
// Construction of SYCL device.
template <typename T, typename detail::enable_if_t<
std::is_same<T, device>::value> * = nullptr>
T make(typename interop<backend::opencl, T>::type Interop) {
return make_device(detail::pi::cast<pi_native_handle>(Interop));
}
// Construction of SYCL context.
template <typename T, typename detail::enable_if_t<
std::is_same<T, context>::value> * = nullptr>
T make(typename interop<backend::opencl, T>::type Interop) {
return make_context(detail::pi::cast<pi_native_handle>(Interop));
}
// Construction of SYCL program.
#ifdef __SYCL_INTERNAL_API
template <typename T, typename detail::enable_if_t<
std::is_same<T, program>::value> * = nullptr>
T make(const context &Context,
typename interop<backend::opencl, T>::type Interop) {
return make_program(Context, detail::pi::cast<pi_native_handle>(Interop));
}
#endif
// Construction of SYCL queue.
template <typename T, typename detail::enable_if_t<
std::is_same<T, queue>::value> * = nullptr>
T make(const context &Context,
typename interop<backend::opencl, T>::type Interop) {
return make_queue(Context, detail::pi::cast<pi_native_handle>(Interop));
}
} // namespace opencl
} // namespace sycl
} // __SYCL_INLINE_NAMESPACE(cl)
| 2,040 |
Subsets and Splits