Search Results

Search found 27144 results on 1086 pages for 'tail call optimization'.

Page 307/1086 | < Previous Page | 303 304 305 306 307 308 309 310 311 312 313 314  | Next Page >

  • Is it possible to connect these machines with ssh tunnels or another mechanism?

    - by frankc
    I am physically on a machine that is behind a firewall and cannot be ssh'd into. Let's call that PC. PC can connect out to two machines, one on AWS and one that is part of a vpn connection, call them A and V, respectively. A and V cannot talk to each other directly. I need to frequently move files between A and V and have been doing this by scp'ing to PC and then scp'ing from pc. Is there a better way to do this? A and V are linux but PC has cygwin.

    Read the article

  • Barnyard Service - MySQL Error

    - by SLYN
    I installed barnyard2 and saved as a service. When I run service barnyard2 start, Barnyard2 is failed. After I run tail -100 /var/log/messages and I encounter a fault like this. ERROR database: 'mysql' support is not compiled into this build of snort#012 Aug 22 11:52:06 barnyard2[25771]: FATAL ERROR: If this build of barnyard2 was obtained as a binary distribution (e.g., rpm,#012or Windows), then check for alternate builds that contains the necessary#012'mysql' support.#012#012If this build of barnyard2 was compiled by you, then re-run the#012the ./configure script using the '--with-mysql' switch.#012For non-standard installations of a database, the '--with-mysql=DIR'#012syntax may need to be used to specify the base directory of the DB install.#012#012See the database documentation for cursory details (doc/README.database).#012and the URL to the most recent database plugin documentation. Aug 22 11:52:06 barnyard2[25771]: Barnyard2 exiting What sould I do for solving this problem? When I installed Barnyard2, I used these commands: # ./configure --with-mysql --with-mysql-libraries=/usr/lib64/mysql # make ; make install (My System is CentOS 6.5 x86_64.)

    Read the article

  • Sharing a printer on Windows home network

    - by Jeff
    I have a home network, and I'm trying to share a printer connected to one of the PCs. The wireless modem through which the internet comes through doesn't reach the entire house, so we have another router set up as well. The modem is 192.168.1.1 (let's call this A) and the router is set up as 192.168.1.100 (let's call this B). Subnet masks 255.255.255.0. The printer is hooked up to a PC on B. If my laptop is connected wirelessly to B, it can see the printer. However, if it's connected to A, it doesn't show up in My Network Places. How can I get it so that I can print to the printer regardless of whether the

    Read the article

  • Sharing a printer on Windows home network

    - by Jeff
    I have a home network, and I'm trying to share a printer connected to one of the PCs. The wireless modem through which the internet comes through doesn't reach the entire house, so we have another router set up as well. The modem is 192.168.1.1 (let's call this A) and the router is set up as 192.168.1.100 (let's call this B). Subnet masks 255.255.255.0. The printer is hooked up to a PC on B. If my laptop is connected wirelessly to B, it can see the printer. However, if it's connected to A, it doesn't show up in My Network Places. How can I get it so that I can print to the printer regardless of whether the

    Read the article

  • Do You Think That SEO Really Effects Your Business?

    Before discussing about the importance of SEO we should know what it is actually. SEO or Search Engine Optimization is actually the procedure of improving the level or excellence of the traffic in a website from various search-engines by the means of natural or unpaid search results as contrasting to the search engine marketing or SEM which actually deals with the paid enclosure. The main importance of the SEO is that it generally makes the website appear higher or earlier in the list of the search results which attracts plenty of visitors from search engines. SEO generally targets different kinds of searches.

    Read the article

  • My feenix nascita mouse isn't moving my cursor. Ubuntu version appears to not matter. How do I debug this, and get the mouse working?

    - by NullVoxPopuli
    I had 13.04, upgraded to 13.10, and did a fresh install of 13.10, no luck. Any ideas? Just tail -f'd my syslog and plugged in the mouse. here is what I got: Oct 26 16:15:50 Orithyia kernel: [83369.618365] usb 1-1.5.2: new full-speed USB device number 6 using ehci-pci Oct 26 16:15:50 Orithyia kernel: [83369.718913] usb 1-1.5.2: New USB device found, idVendor=04d9, idProduct=a081 Oct 26 16:15:50 Orithyia kernel: [83369.718919] usb 1-1.5.2: New USB device strings: Mfr=1, Product=2, SerialNumber=0 Oct 26 16:15:50 Orithyia kernel: [83369.718921] usb 1-1.5.2: Product: USB Gaming Mouse Oct 26 16:15:50 Orithyia kernel: [83369.718924] usb 1-1.5.2: Manufacturer: Holtek Oct 26 16:15:50 Orithyia kernel: [83369.722486] input: Holtek USB Gaming Mouse as /devices/pci0000:00/0000:00:1a.0/usb1/1-1/1-1.5/1-1.5.2/1-1.5.2:1.0/input/input13 Oct 26 16:15:50 Orithyia kernel: [83369.722786] hid-generic 0003:04D9:A081.0004: input,hidraw3: USB HID v1.10 Keyboard [Holtek USB Gaming Mouse] on usb-0000:00:1a.0-1.5.2/input0 Oct 26 16:15:50 Orithyia kernel: [83369.729362] hid-generic 0003:04D9:A081.0005: usage index exceeded Oct 26 16:15:50 Orithyia kernel: [83369.729366] hid-generic 0003:04D9:A081.0005: item 0 2 2 2 parsing failed Oct 26 16:15:50 Orithyia kernel: [83369.729379] hid-generic: probe of 0003:04D9:A081.0005 failed with error -22 Oct 26 16:15:50 Orithyia kernel: [83369.731759] hid-generic 0003:04D9:A081.0006: hiddev0,hidraw4: USB HID v1.10 Device [Holtek USB Gaming Mouse] on usb-0000:00:1a.0-1.5.2/input2 Oct 26 16:15:50 Orithyia mtp-probe: checking bus 1, device 6: "/sys/devices/pci0000:00/0000:00:1a.0/usb1/1-1/1-1.5/1-1.5.2" Oct 26 16:15:50 Orithyia mtp-probe: bus: 1, device: 6 was not an MTP device So, I see it did see it is a mouse... but the cursor doesn't move when I move the mouse. Not sure where to debug from here.

    Read the article

  • What effect does using itemprop="significantLinks" on anchors have for SEO and what is the ideal use?

    - by hdavis84
    I'm practicing application of microdata via http://schema.org. Anyone who's browsed the documentation there knows that there's a lot of need for improvement for more clear understandings on use for each property. My question on this post is more about the "significantLinks" property and how it effects SEO for on page, in content anchored text. Does anyone have any more information regarding whether its good to use for link optimization? I understand what schema.org means that it's to be used on "non-navigational links" and those links should be relevant to the current page's meaning. But will using this property hurt SEO or make SEO better for each page?

    Read the article

  • searchd under runit continues writing to the runit's log

    - by Eugene
    searchd (Sphinx) run file: #!/bin/sh set -e APP_PATH=/srv/application TARGET_USER=user exec chpst -u $TARGET_USER /usr/bin/searchd --pidfile --nodetach --config $APP_PATH/current/config/production.sphinx.conf tail /var/log/sphinx/current 2014-06-07_18:13:56.87885 precached 9 indexes in 0.497 sec 2014-06-07_18:13:57.13740 precached 9 indexes in 0.497 sec 2014-06-07_18:13:57.88113 precached 9 indexes in 0.497 sec 2014-06-07_18:13:57.89167 precached 9 indexes in 0.497 sec 2014-06-07_18:13:59.75555 precached 9 indexes in 0.497 sec 2014-06-07_18:13:59.81554 precached 9 indexes in 0.497 sec 2014-06-07_18:14:00.33466 precached 9 indexes in 0.497 sec ... it continues to write the same line until sv stop sphinx ... Everything works fine, seachd starts and responds to the queries. But how to make logs to be less repetitive? When I start Sphinx manually it prints the "precached 9 indexes" just once.

    Read the article

  • Access denied to EFS encrypted files after PC joins domain

    - by mjmarsh
    I'm experiencing strange behavior with Windows Encrypted File System: I have a machine that is in workgroup mode (not joined to a domain) I encrypt an entire directory structure on the machine (basically a folder and subfolders with data files for my application). My application writes and reads files from the encrypted file hierarchy as a local Windows user (let's call the account 'SecureUser'). This works fine I then join the PC to a domain (Let's call it 'TEST') Afterwards, processes running as the local 'SecureUser' account can't read the files it wrote originally when it was off the domain (What is also strange is that the files are listed as "read only" now and I cannot unset this flag via Windows Explorer or the command line, even though it looks like it succeeds) I then 'un-join' the PC from the domain and everything works again Is there something about changing domain membership on a PC that changes the behavior of EFS so that previously encrypted files cannot be read, even by the originating user? Thanks in advance

    Read the article

  • What exactly are Link Relation Values?

    - by bckpwrld
    From REST in Practice: Hypermedia and Systems Architecture: For computer-to-computer interactions, we advertise protocol information by embedding links in representations, much as we do with the human Web. To describe a link's purpose, we annotate it. Annotations indicate what the linked resource means to the current resource: “status of your coffee order” “payment” and so on. We call such annotated links hypermedia controls, reflecting their enhanced capabilities over raw URIs. ... link relation values, which describe the roles of linked resources ... Link relation values help consumers understand why they might want to activate a hypermedia control. They do so by indicating the role of the linked resource in the context of the current representation. I interpret the above quotes as saying that Hypermedia control contains both a link to a resource and an annotation describing the role of linked resource in the context of the current representation. And we call this annotation ( which describes the role of linked resource ) a link relation value. Is my assumption correct or does the term link relation value actually describe something different? Thank you

    Read the article

  • Many user stories share the same technical tasks: what to do?

    - by d3prok
    A little introduction to my case: As part of a bigger product, my team is asked to realize a small IDE for a DSL. The user of this product will be able to make function calls in the code and we are also asked to provide some useful function libraries. The team, together with the PO, put on the wall a certain number of user stories regarding the various libraries for the IDE user. When estimating the first of those stories, the team decided that the function call mechanism would have been an engaging but not completely obvious task, so the estimate for that user story raised up from a simple 3 to a more dangerous 5. Coming to the problem: The team then moved to the user stories regarding the other libraries, actually 10 stories, and added those 2 points of "function call mechanism" thing to each of those user story. This immediately raised up the total points for the product of 20 points! Everyone in the team knows that each user story could be picked up by the PO for the next iteration at any time, so we shouldn't isolate that part in one user story, but those 20 points feel so awfully unrealistic! I've proposed a solution, but I'm absolutely not satisfied: We created a "Design story" and put those annoying 2 points over it. However when we came to realize and demonstrate it to our customers, we were unable to show something really valuable for them about that story! Here the problem is whether we should ignore the principle of having isolated user stories (without any dependency between them). What would you do, or even better what have you done, in situations like this? (a small foot-note: following a suggestion I've moved this question from stackoverflow)

    Read the article

  • Excel Countif external date

    - by Duall
    I am making an Excel 2010 spreadsheet to log support calls, services, and installations that each member would fill out. Due to being paid by job rather than by hour, there is a need for it to count each of these ("Call", "Service", "Install") there is in any given time span. The entry of the data itself would be in Sheet 1, and then a 'splash screen' of sorts would be in Sheet 2. Here I would like to be able to put a date range and it would display how many of each there is. I already can do the COUNTIF statement, =COUNTIF(Activity!$B:$B,"Call") but I don't know how to: a) Add in the prerequisite for a date so it only gets "Calls" within a certain time frame. b) Take the date it looks for from a cell on the splash screen.

    Read the article

  • How to persuade C fanatics to work on my C++ open source project?

    - by paperjam
    I am launching an open-source project into a space where a lot of the development is still done Linux-kernel-style, i.e. C-language with a low-level mindset. There are multiple benefits to C++ in our space but I fear those used to working in C will be scared off. How can I make the case for the benefits of C++? Specifically, the following C++ attributes are very valuable: concept of objects and reference-counting pointers - really don't want to have to malloc(sizeof(X)) or memcpy() structs templates for specialising whole bodies of code with specific performance optimizations and for avoiding duplication of code. template metaprogramming related to the above syntactic sweetness available (e.g. operator overloading, to be used in very small doses) STL Boost libraries Many of the knee-jerk negative reactions to C++ are illfounded. Performance does not suffer: modern compilers can flatten dozens of call stack levels and avoid bloat through wide use of template specializations. Granted, when using metaprogramming and building multiple specializations of a large call tree, compile time is slower but there are ways to mitigate this. How can I sell C++?

    Read the article

  • Moving files with batch files from one pc to a server, to a another pc - worried about disk corruption

    - by AnchientAnt
    I use scheduled tasks that calls a batch file, that calls more batch files to move about three files from a pc, to a server, then to multiple other pcs. It all happens very quickly, as they are small files. Are there any pitfalls for how fast these transfers happen? I'm just mildly concerned about causing some disk corruption somehow. I use logic like 1. Call MapToPc if files exist then move file to folder on server. Disconnect 2. Call SendtoPCs If files exist (the files just moved to the server) then MapToPCs Move all files Disconnect All of this happens in about 2 secs or less. edit: this on windows 7, server 2003, xp respectively

    Read the article

  • Redirecting HTTP traffic from a local server on the web

    - by MrJackV
    Here is the situation: I have a webserver (let's call it C1) that is running an apache/php server and it is port forwarded so that I can access it anywhere. However there is another computer within the webserver LAN that has a apache server too (let's call it C2). I cannot change the port forwarding nor I can change the apache server (a.k.a. install custom modules). My question is: is there a way to access C2 within a directory of C1? (e.g. going to www.website.org/random_dir will allow me to browse the root of C2 apache server.) I am trying to change as little as possible of the config/other (e.g. activating modules etc.) Is there a possible solution? Thanks in advance.

    Read the article

  • What is the/Is there a right way to tell management that our code sucks?

    - by Azkar
    Our code is bad. It might not have always been considered bad, but it is bad and is only going downhill. I started fresh out of college less than a year ago, and many of the things in our code puzzle me beyond belief. At first I figured that as the new guy I should keep my mouth shut until I learned a little more about our code base, but I've seen plenty to know that it's bad. Some of the highlights: We still use frames (try getting something out of a querystring, almost impossible) VBScript Source Safe We 'use' .NET - by that I mean we have .net wrappers that call COM DLLs making it almost impossible to debug easily Everything is basically one giant function Code is not maintainable. Each page has multiple files that are created every time a new page is made. The main page basically does Response.Write() a bunch of times to render the HTML (runat="server"? no way). After that there can be a lot of logic on the client side (VBScript), and finally the page submits to itself (often time storing many things in hidden fields) where it then posts to a processing page which can do things such as save the data to the database. The specifications we get are laughable. Often times they call for things like "auto-populate field X with either field Y or field Z" with no indication of when to choose field Y or field Z. I'm sure some of this is a result of not being employed at a software company, but I feel as if people writing software should at least care about the quality of their code. I can't even imagine that if I were to bring something up that anything would be done soon, as there is a large deadline looming, but we are continuing to write bad code and use bad practices. What can I do? How do I even bring these issues up? 75% of my team agree with me and have brought up these issues in the past, yet nothing gets changed.

    Read the article

  • What is needed to become a skillful and professional software developer ?

    - by silentbang
    I am interested in studying algorithms and my strong points are problem-solving and logical thinking. However what I am guided toward is web-developing languages, SQL and they seem to not exploit the best out of me. I think it's because lots of people can learn these things easily, even steal ideas and copy code. Is that wrong? I was told that "learning algorithms is just for optimization, so you just need to know it (not a deep understanding)", but I think algorithms are my savior; it differentiates one man's abilities from others'. Also I'm learning C++ to benefit my future career. Many people say that web developer is a future trend, which worries me. I don't know what I need to standout in my career, and should I balance between web and software developing or just one? Should I work hard on algorithms?

    Read the article

  • Why am I getting [mount error(22): Invalid argument] while trying to mount SMB network drive?

    - by Steve_
    Disclaimer: I am very new to Linux :) Anyway, onward: I have a fresh instance of Ubuntu Server (12.04.1 LTS) running on my network and I want to mount a network drive to the server so I can access the contents. The network drive is a SAMBA compatible drive running Darwin OS. If I run the following command: smbclient -L //192.168.0.2 -U myuser It prompts me for the password and then displays output similar to: Domain=[SERVER01] OS=[Darwin] Server=[@(#)PROGRAM:smbd PROJECT:smbx-105.4.0] Sharename Type Comment --------- ---- ------- Comp Staff's Public Folder Disk CompRaid03 Disk Dropbox Disk Groups Disk IPC$ IPC Public Disk Users Disk compstaff Disk However, when I try and mount the CompRaid03 share, using this command: sudo mount -t cifs //192.168.0.2/CompRaid03 /mnt/myshare -o username=myuser I get the same password prompt, but after putting the correct password in, I received this error: mount error(22): Invalid argument dmesg | tail returns: [23576.037373] CIFS VFS: cifs_mount failed w/return code = -22 I don't understand what is wrong with this command. I've managed to mount a share on my current (Windows 8) machine using basically the same command but with a different IP address and share name (obviously). I've spent a good few hours trying to solve this and got no where. Any help or pointers would be greatly appreciated. Thanks Steve EDIT As suggested I've also trued using "user=" instead of "username=": sudo mount -t cifs //192.168.0.2/CompRaid03 /mnt/svnrepo -o user=myuser This results in the same "Invalid argument" error.

    Read the article

  • Multi-monitors and the corners of the screen

    - by Neil Barnwell
    I currently have two monitors (let's call them "left" and "middle" - you'll see why in a sec), and would love three (let's call it "right"). However, I often will throw the mouse-cursor up to the top-right corner on "middle" because for a maximised window that's where the close button will be. If I add the "right" monitor, the mouse cursor will just carry on off "middle" into "right", and aiming for the close button on "middle" will become more difficult. I'm currently using UltraMon 3.10, but does anyone know of a way to get the mouse to stick to the corners, so that it doesn't go off into the other monitor (Synergy allows this by configuring the "gap" where the mouse is allowed to travel from monitor to monitor).

    Read the article

  • software to view logs on remote server

    - by Tarun
    I have configured syslog server on my linux machine like this Linux Machine (syslog client) -- Linux Machine (syslog server) I have configured it and its working properly. Now, the problem is I want to look at the logs located on remote server in a particular directory for example /var/log/host-ip/httpd.log. One way is that I use ssh for that and then see those logs using commands like tail etc. So is there any Browser based graphical utility which I can use on my machine in which I just give the remote folder location, ip-address etc. and the software shows me the logs present on that remote server. Thanks.

    Read the article

  • Need some critique on .NET/WCF SOA architecture plan

    - by user998101
    I am working on a refactoring of some services and would appreciate some critique on my general approach. I am working with three back-end data systems and need to expose an authenticated front-end API over http binding, JSON, and REST for internal apps as well as 3rd party integration. I've got a rough idea below that's a hybrid of what I have and where I intend to wind up. I intend to build guidance extensions to support this architecture so that devs can build this out quickly. Here's the current idea for our structure: Front-end WCF routing service (spread across multiple IIS servers via hardware load balancer) Load balancing of services behind routing is handled within routing service, probably round-robin One of the services will be a token Multiple bindings per-service exposed to address JSON, REST, and whatever else comes up later All in/out is handled via POCO DTOs Use unity to scan for what services are available and expose them The front-end services behind the routing service do nothing more than expose the API and do conversion of DTO<-Entity Unity inject service implementation to allow mocking automapper for DTO/Entity conversion Invoke WF services where response required immediately Queue to ESB for async WF -- ESB will invoke WF later Business logic WF layer Expose same api as front-end services Implement business logic Wrap transaction context where needed Call out to composite/atomic services Composite/Atomic Services Exposed as WCF One service per back-end system Standard atomic CRUD operations plus composite operations Supports transaction context The questions I have are: Are the separation of concerns outlined above beneficial? Current thought is each layer below is its own project, except the backend stuff, where each system gets one project. The project has a servicehost and all the services are under a services folder. Interfaces live in a separate project at each layer. DTO and Entities are in two separate projects under a shared folder. I am currently planning to build dedicated services for shared functionality such as logging and overload things like tracelistener to call those services. Is this a valid approach? Any other suggestions/comments?

    Read the article

  • SEO Tips - Updating Your Content and Avoiding Duplicate Filters

    If you are just getting into Search Engine Optimization, it's important that you are aware that content is the most important thing in your website in order to get high ratings. If you decide to hire an SEO Consultant, make sure that you explain to them exactly what you want your website to look like and the content that you would like it to feature. Having fresh and relevant content in your website is the key as that will bring web crawlers back frequently. There are different ways of achieving this while avoiding getting your website removed off of a search engine due to duplicate content.

    Read the article

  • Boost Your SEO Efforts With Twitter

    Search engine optimization techniques continue to expand so that more innovative ways of drawing traffic to your website can be used effectively. While the online revolution has changed many things in a common man's life, one exciting facet of the World Wide Web is social networking. With social networking giants such as Twitter, Facebook, LinkedIn, Orkut and MySpace hitting the pinnacle of popularity, people have the luxury of making online friends, interact with people irrespective of national boundaries and discuss world issues without any ethnic or racial biases. Webmasters and social network gurus have also capitalized on this immense opportunity to expand their businesses and draw more traffic to their websites.

    Read the article

  • Naming the implementation version of an interface function

    - by bolov
    When I need to write an implementation version of an interface function, I put the implementation function within a impl namespace, but with the same name as the interface function. Is this a bad practice? (the same name part, the namespace part I am confident it’s more than OK). For me, who I write the code, there is no confusion between the two, but I want to make sure this isn’t confusing for someone else. One other option would be to append impl suffix to the function name, but since it is already in a separate namespace named impl it seems redundant. Is there an idiomatic way to do this? E.g.: namespace n { namespace impl { // implementation function (hidden from users) // same name, is it ok? void foo() { // ... //sometimes it needs to call recursively or to call overloads of the interface version: foo(); // calls the implementation version. Is this confusing? n::foo(); // calls the interface version. Is this confusing? // ... } // namespace impl // interface function (exposed to users) void foo() { impl::foo(); } } // namespace n

    Read the article

  • how to recover data from my disk - I accidentally copied with dd an iso on it

    - by sijoune
    I wanted to create a bootable usb from an iso image and i accidentally put as the output of the dd, instead of my usb drive, one of my hard disks. The iso was 3,3 GB and my disk is 1TB! And it was almost full. Can i at least restore the data that has not been overwritten? Right now i can't even mount it. I get this error: Error mounting /dev/sdd1 at /media/main/UDF Volume: Command-line `mount -t "udf" -o "uhelper=udisks2,nodev,nosuid,uid=1000,gid=1000,iocharset=utf8,umask=0077" "/dev/sdd1" "/media/main/UDF Volume"' exited with non-zero exit status 32: mount: wrong fs type, bad option, bad superblock on /dev/sdd1, missing codepage or helper program, or other error In some cases useful info is found in syslog - try dmesg | tail or so Also since i know which filesystem my disk used if i reformat it to this filesystem is there any chance i can mount it and retrieve the rest of the files?

    Read the article

< Previous Page | 303 304 305 306 307 308 309 310 311 312 313 314  | Next Page >