Search Results

Search found 5067 results on 203 pages for 'squid deb proxy'.

Page 141/203 | < Previous Page | 137 138 139 140 141 142 143 144 145 146 147 148  | Next Page >

  • Where to start building a BaaS

    - by Wesley
    I'm building a Cloud Platform, and the next phase of design involves building an extensible BaaS back end. (see http://youtu.be/lNi-05-PyEw) The reason I think we can attempt this, is there are dozens of these kinds of extensible back end data proxy's popping up almost daily at this point, which tells me the enabling technology is there to build one from scratch in a few months. I'd like to start in the right area: What kind of Dev background should I look for? What kind of tech stack should I build on? What kind of costs can I expect in terms of man-hours, etc... I know there isn't one right answer here, but I think this is the right sub to post this in, and credit will go towards to most constructive answer.

    Read the article

  • Mehr Sicherheit für Netzwerkverbindungen

    - by DBA Community
    Der Zugriff auf Datenbanken über das Netzwerk stellt aus Security Sicht einen ausgesprochen kritischen Vorgang dar, der unbedingt vor Missbrauch geschützt werden muss. Deshalb ist es auch nicht weiter verwunderlich, dass im Security Ecosystem etliche Produkte angeboten werden, die diesen Zugriff sichern helfen: Das beginnt bei Firewalls mit SQL Net Proxy, geht über Produkte wie die Oracle Database Firewall, die einen Schutz vor SQL Injection Angriffen über das Netzwerk leisten, und endet etwa bei den Angeboten zur Netzwerkverschlüsselung, wie sie im Oracle Datenbankumfeld vor allem die Advanced Security Option anbietet. Aber vor jedem Einsatz schwieriger oder kostspieliger Mittel zur Steigerung der Sicherheit einer Datenbank steht der Einsatz solcher Mittel, die ohne zusätzliche Kosten oder relativ einfach zu implementieren sind. Dazu gehören das bereits in einem Community Artikel andiskutierte  Härten der Datenbank oder das in einem weiteren Artikel angesprochene Umsetzen des  Prinzips des least privilege. Im vorliegenden Artikel soll darauf eingegangen werden, wie die Verbindungsaufnahme zur Datenbank über einen Listener Prozess sowie die Netzwerkverbindung zwischen Client und Datenbank über SQL Net eigene Mittel so konfiguriert werden können, dass dies die Sicherheit einer Datenbank ohne Zusatzkosten erhöht. Weiter zum Tipp.

    Read the article

  • Assign subdomains to seperate ports on web server

    - by Michael Frank
    I have set up an Abyss web server as a little experiment, and I want to know if it is possible to assign subdomains to different ports on the machine the web server is running on. I have a couple webUIs that I'd like to assign subdomains: 192.168.1.1:8000 becomes example.com/webui1/ 192.168.1.1:8001 becomes example.com/webui2/ The webUIs are available by accessing their ports via example.com:8000. I have tried using a reverse proxy, but it seems that this is only usable on one internal IP at a time. What other options do I have?

    Read the article

  • How to sell or rent your house on Craigslist

    Im in the process of relocating. Im relocating to a state Im not familiar with and have no close friends/family that can serve as my go-to proxy on a real-time basis for house hunting help. Short of being there, this leaves the internet for my wife and I to do searching and narrow things down. We had 2 trips to the new location to do some house hunting. This means that we had to do all our research before to make those 2-day adventures worthwhile. Welcome to the internet and real estate. Now, there...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • ValueError: too many values to unpack in a tuple

    - by falosi
    Please put some light on why am getting a too many to unpack (ValueError in my for loop).Have tried deb naislist = [('CONTROL FILE', '0', '0', '0'), ('REDO LOG', '0', '0', '0'), ('ARCHIVED LOG', '.69', '.59', '3'), ('BACKUP PIECE', '46.54', '0', '192'), ('IMAGE COPY', '0', '0', '0'), ('FLASHBACK LOG', '10.15', '6.31', '82'), ('FOREIGN ARCHIVED LOG', '0', '0', '0')] print "size of naislist is ",len((naislist)) heading = ('MAIN MENU', 'LEVELS', 'LEVEL2', 'LEVEL3') rearrange = dict(zip((0, 1, 2, 3), (len(str(x)) for x in heading))) for tu, x in naislist: rearrange.update((i, max(rearrange[i], len(str(el)))) for i, el in enumerate(tu)) rearrange[4] = max(rearrange[4], len(str(x))) forkit = '|'. join('%%-%ss' % rearrange[i] for i in xrange(0, 4)) print '\n'.join((forkit % heading, '-|-'.join(rearrange[i] * '-' for i in xrange(4)), '\n'.join(forkit % (a, b, c, d) for (a, b, c), d in naislist)))

    Read the article

  • NDIS Driver Filter VS API Hooking

    - by Smarty Twiti
    I've seen many developers asking for "How to intercept in/out HTTP packets ", "How to modify them on the fly". The most "clean" answer I've seen is to make a kernel-mode-driver filter from the scratch (TDI for XP and earlier winx9 or NDIS for NT systems). An other way, is to use a user-mode-driver like Windivert, also Komodia has a great solution (without writing any single code). The idea behind this introduction is just I want to know is API Hooking can be considered as alternative of writing of whole of driver-filter? writing a driver from the scratch is not an easy task, why just not Hooking the HttpSendRequest or any other API used by the browser? There are many free/commercial libraries to do this in a safe manner (eg: EasyHook, Mhook, Nektra..). I'm not the first who ask, there already Sockscap that uses Hook(DLL injection) to change behavior to other applications and force them to use a Socks proxy, also Form grabbing attack 'used by keylogger..

    Read the article

  • Disabling LDAP Signing on Windows PDC in Local Policy

    - by Golmaal
    I just tripped over my own feet it seems. Playing around on a Windows 2008 R2 server (set up as domain controller), I was intrigued by certain warning event (event id 2886) which says: "To enhance the security of directory servers, you can configure both Active Directory Domain Services (AD DS) and Active Directory Lightweight Directory Services (AD LDS) to require signed Lightweight Directory Access Protocol (LDAP) binds." So I thoughtlessly did some Googling and set the relevant policies which enforce LDAP signing. Now I don't remember but I may have done that using Local Policy. Now I have setup a pfsense box which must authenticate AD users via LDAP. While the firewall can communicate over secure channel, it is difficult to manage the same for other packages such as Squid and SquidGuard. So now I have to disable i.e. undo those policy changes. The problem is that they are greyed out! The policies in question are LDAP server signing and LDAP client signing. I don't remember what I did but when I access these policies from Local Policy editor on the server, they are set to "Require Signing" and are greyed out. The same policies can still be set via Default Domain Controller option in Group Policy editor. So how can I reset these greyed out policies? Thanks

    Read the article

  • About cdn architecture to route way

    - by Tony Lee
    Our web system, use the third-party cdn service. Assume that the user set the local dns with the googledns or opendns to visit our web sites, so cdn service will select the closest cdn proxy node. all right, but in fact the user's actual access position might outside there, cdn service may chose the one furthest away from the user node, so static resource access slower.. At present, my idea is if user local set dns server with googledns, and then first one we get the actual ip address of the user, tracerote to test a best routing lines, set up a cookie in user browser, and then set 302 header for response to jump to the which best cdn node. Whether the user's browser side traceroute tool can provide the best route decision-making ? Because we find that, once the user to set local dns server with the foreign network segment, for example : set dns with 8.8.8.8, so cdn routing will choose the foreign service node.

    Read the article

  • Getting Started Quickly

    - by Owen Allen
    If you're interested in using Ops Center, you'll want to get up and running as quickly and effectively as possible. One way to do this would be to work your way through the documentation library - use the Linux or Oracle Solaris install guides, then go through the Feature Guide and Admin Guide to start using the software. They're thorough, but they're a lot of reading. But if you're looking to install a simple deployment quickly, and you don't want to do all of the configuration work right off the bat, you can use the Quick Start Guide. It's a streamlined procedure that runs you through installing a single Enterprise Controller and co-located Proxy Controller, and then shows you how to discover assets quickly. Once you've discovered these assets, it describes how to use the analytics feature to view their performance, and use monitoring to keep track of their statuses and health. You'll have to do some additional configuration to use features like OS provisioning, OS updates, and virtualization, but the Quick Start guide gives you an overview of how to install and start using features quickly.

    Read the article

  • package update in ubuntu on vitual box

    - by Ayusman
    while following one of the tutorials I needed to install a curl package. Here are my system configuration: Virtualbox 4.1 on Windows 7 64 bit. I have installed ubuntu 12.04 on the virtual box. When I try the following: sudo apt-get install curl I see the following errors: [sudo] password for userMe: Reading package lists... Done Building dependency tree Reading state information... Done Package curl is not available, but is referred to by another package. This may mean that the package is missing, has been obsoleted, or is only available from another source E: Package 'curl' has no installation candidate userMe@userMe-ubuntu:~$ When I try the same on a direct installation of ubuntu on a laptop (not a virtual machine) there is no problem. I am assuming this may be because of network settings. Please note that my work PC is in a corporate network and uses auto proxy settings (using a .pac file) for internet access etc. Any help will be appreciated.

    Read the article

  • Comparing the Performance of Visual Studio&apos;s Web Reference to a Custom Class

    As developers, we all make assumptions when programming. Perhaps the biggest assumption we make is that those libraries and tools that ship with the .NET Framework are the best way to accomplish a given task. For example, most developers assume that using <a href="http://www.4guysfromrolla.com/articles/120705-1.aspx">ASP.NET's Membership system</a> is the best way to manage user accounts in a website (rather than rolling your own user account store). Similarly, creating a Web Reference to communicate with a <a href="http://www.4guysfromrolla.com/articles/100803-1.aspx">web service</a> generates markup that auto-creates a <i>proxy class</i>, which handles the low-level details of invoking the web service, serializing parameters,

    Read the article

  • Need advise for choosing software\hardware for virtualization.

    - by Anatoly
    Currently we have these servers : Windows SBS 2003 premium on IBM X266 double Xeon F43, 2GB ram. DC, exchange (70 users), Mssql. Windows 2003 R2 32bit on IBM x3400 with double XEON E5310 and 4GB ram. Terminal server (40+ users), ERP application based on uniPaaS platform from Magicsoftware, and Pervasive sql. Ubuntu 8.04 (simple pc box) with squid proxy, GLPI system and PHPBB3 forum for internal use. Recently number of concurrent users on Terminal server passed 40 users in rush hours and it gets stuck frequently. Therefore we need an upgrade. I think about transfer all physical servers to virtual servers based on cluster of 2 physical servers for reducing downtime. I think we will grow till 50-60 concurrent terminal users in rush hours. I also plan to virtualize 10-15 Win XP/7 workstation (office,ERP etc), and there is a little probability for Asterisk\Hylafax for 100 users (if it possible on same VM). Also we need NAS storage for 2-3TB. What hardware upgrade/purchase we need for complete this task? Which VM solution is preferable VmWare or Hyper-V? What backup software should we choose? Acronis or something another? Thank you in advance.

    Read the article

  • Ubuntu 12.04 PXELINUX does not boot RHEL Kernel and Initrd

    - by utpal
    I have successfully setup PXE server on Ubuntu 12.04 with DNSMASQ for DHCP Proxy Service, TFTPD-HPA for TFTP service, NFS-KERNEL-SERVER, APACHE2 and SYSLINUX for pxelinux.0 bootloaded needed for pxe boot using the following POST: http://ubuntuforums.org/showthread.php?t=1606910 I was successfully able to pxe boot a client to a Ubuntu 12.04 LIVE CD. Next, I want to PXE boot a client to a RHEL 6.5 x64 Kernel and initrd image. I dont want to install, just boot a client so that it mounts initrd and I can get a minimal filesystem on the client. How can I do that? Please Help!!

    Read the article

  • WAN Optimization for Small Office/Home Office

    - by TiernanO
    I have been reading up on WAN optimization for the last while, mostly out of interest of speeding up my own internet connections, but also to speed up the office internet connection. At home, I have 2 cable modems plugged into a RouterBoard RB750, which load balances the connections. In the office, we have a single connection into a NetGear router. Most of the WAN Optimization products I have seen, seem to be prohibitively expensive, but also seem to be based on the idea of having multiple branches around the world. What I am looking for, ideally, is as follows: software install: I am "guessing" I need to install it in 2 places: one in the office or house, and one in "the cloud". any connections going to, say, The US (we are in Europe, but our backup's live in the US currently, which would be something important to speed up) would be "tunnelled" though the Optimizer. If downloading or uploading large files, open multiple connections between both "the cloud" and the optimizer... This is where a lot of speed could be gained. finally, for items not compressed, they would be compressed on the cloud side of things, also items that are already on the optimizer could be not sent again. kind of like RSync or Proxy servers... So, is there something that can be done? Is it available using off the shelf components (some magic script with SSH, Squid, Linux and duct tape) or is it something that needs to be purchased? or even an Open Source Project that does 90% of what i am asking?

    Read the article

  • Is there any way to simulate a slow connection between my server and an iPad (without installing anything on the server)?

    - by Clay Nichols
    Some of our webapp users have difficulty on slower connections. I"m trying to get a better idea of what that "speed barrier is" so I'd like to be able to test a variety of connection speeds. I've found ways to do this on Windows but no on the iPad, so I'm looking more for some sort of proxy service that'll work with any device (not running ON that device) I did find an article about using the CharlesProxy and providing a connection to another device, but I was hoping for something simpler (need not be free) Constraints * We are on a shared server so we can't install anything and we are limited in our control over that server. * I'd like to test an iPad, Android Tablet, Windows PC.

    Read the article

  • Another Update to SQL Server Configuration Scripting Utility

    - by Bill Graziano
    I’ve been gradually adding features to my utility that scripts the configuration of a SQL Server.  Since my last post I’ve added the following features: Skip any encrypted object in a database Script alerts, alert notifications and operators Script audits Always script model, master and msdb to capture any user-defined objects in those databases Logins are now scripted so that everything for a login is grouped together. There’s a second section in the logins that handles default databases.  In many cases a login’s default database is a mirror target and can’t be set.  This is now handled gracefully.  It also includes a separate section for all default databases so those can be quickly set in the event of a disaster. Script credentials Script proxy accounts Script database mail My goal is still to get everything outside a database scripted.  This release is enough that I can keep my mirror target servers in sync with their principals.

    Read the article

  • Caching preventing users seeing site updates

    - by Timmeh
    I'm experiencing a caching issue I can't explain. This is happening across browsers, IPs and ISPs. If a user force-refreshes, they see the new content. If they then refresh or return to the page, the old one displays. I've tried using headers via PHP such as header( 'Expires: Sat, 26 Jul 1997 05:00:00 GMT' ); header( 'Last-Modified: ' . gmdate( 'D, d M Y H:i:s' ) . ' GMT' ); header( 'Cache-Control: no-store, no-cache, must-revalidate' ); header( 'Cache-Control: post-check=0, pre-check=0', false ); header( 'Pragma: no-cache' ); Laid out correctly, at the very beginning of the file. The problem persists. A pan-ISP proxy is unlikely. Suggestions?

    Read the article

  • wifi works only after connecting through wire

    - by orustam
    I have fresh installed ubuntu 12.04. it is my first ubuntu installation and i'm a bit confused about the network connection. Wifi shows up and connects(at least it shows that the connection is established), but i can't open any pages, i've tried to ping some sites and it fails either. If i try to connect through a wire it works, what is interesting to me is that after i used my wire connection i can use my wifi properly without a wire pluged in. i think it probably has to do with my settings? I tried to find a solution but can figure it out on my own. My Proxy set to none(have applied it system wide) Please help me if you have any clue:)

    Read the article

  • Alternatives to using cookies?

    - by theclueless1
    Whate are alternatives to using cookies/client-side storage for a PHP/MySQL based site on Apache. Scenario/Requirements: I want to try using some anti-bot code to prevent specific scrapers etc. from accessing the site. I would like to run this code before launching the rest of the site (before DB access etc.). I don't want to constantly run the same code on every page-load after a visitor has passed the initial check. I'd like to avoid the use of Cookies/Client side storage if at all possible. The only solution I can currently think of is to write files to the server based on the visitors IP/UA, or to write a list of them to a single file. Yet this has the limitation of multiple users through a proxy/same connection, etc ... So, any ideas/suggestions? Or am I simply over working the issue?

    Read the article

  • Transfer domain from godaddy to another registrar and (somewhat) maintain privacy?

    - by anonymous-one
    Does anyone know if its possible in any ways to do this? Transfer a private domain (via domainsbyproxy) registered at godaddy to another registrar. While I know moving a domain from godaddy requires the privacy to be disabled, what we were thinking of doing is the following: a) Update the whois info inside domains by proxy to a 3rd party. Most likely a law firm who would in essnece become our 'custodian'. b) Wait 60 days (?) c) Disable domain privacy. Now the law firms details would become public. d) Conduct the domain transfer. e) Take ownership of the domain at the new registrar. Does something like this sound possible? Does anyone know any other way to handle this? Thanks.

    Read the article

  • screen blacks out after pressing a key on keyboard, specially enter key

    - by ahmad598
    I have natty, and i have installed kubuntu-desktop package. before installing it, everything was fine but after that, this thing happens: when I'm typing, after stroking a key, screen blacks out, fan speeds up, and i have to turn off machine by pressing power button. it just happens in many different apps (chromium, hotot, kde proxy settings, ...) and in every environment: unity, kde, gnome 2. i don't know what details i must provide with this question, and i don't like to remove my kde desktop, or re-install my ubuntu. any ideas are appreciated ;-)

    Read the article

  • 2 Computers, same network, different outgoing speeds when uploading to internet?

    - by user117339
    I have 2 work machines in my office, a PowerMac G5 and a MacBook Air. Both behind an IPCop firewall. The PowerMac is connected through a gigabit switch, the MacBook Air is connected through a Netgear 802.11g access point that is then plugged into the gigabit switch. There is also a FreeNAS box, both machines are able to read and write files to it at close to their pipe speeds. The main problem is when I am trying to upload files to the internet at large. The G5 is only hitting 0.1 - 0.25 Mbps. The Macbook is able to hit 2-3 Mbps. The setup (G5 / IPCop / Network) has been the same for 5 years. The issues with the internet speed started about 3 months ago. I hadn't tested on the Macbook at this point. I had complained to the ISP, they said their modem needed a firmware update, did that nothing changed. Reset IPCop, turned off squid, etc. No changes. The ISP switched the office over to a better plan with a theoretical 6 Mbps up, still no change. At this point I tried testing the Macbook, and lo and behold there's the speed. But why? I have tried changing out everything, cables, switches, using another ethernet port on the G5, wiping the system, using DHCP, using manual IPs, changing DNS servers, etc. Nothing works. I figured that if there was something horribly wrong with the network, then internally I would find a similar issue, but that is perfect. iperf, ping, etc show no dropped packets and near saturation of the internal network. I'm at a loss as to what the heck is going on. Any ideas would be appreciated! Below are some screenshots of speedtest.net: G5: Macbook Air:

    Read the article

  • I wan't to make PC for library. And have some problem ))

    - by Doroff
    I use Ubuntu 12.04. For make .desktop I used this instructions: http://www.instructables.com/id/Setting-Up-Ubuntu-as-a-Kiosk-Web-Appliance/step4/Set-up-Chromium/ 1 problem: No users can't download kiosk.desktop - they download ubuntu.desktop and change that properties in home/user/.dmrc . How can I fix that problem? Once I put all properties that I maked for kiosk.desktop into ubuntu.desktop and it's start work...but on every created users, and after I reinstalled system. 2 problem: Can I write in .desktop which program users can use? If yes-how? 3 problem: Which programm is better to use in proxy for Ubuntu 12.04? Sorry for my english and thanks Yuri

    Read the article

  • Managing service references and endpoint configurations for Silverlight applications

    Youve written your service. Youve written your Silverlight application. You Add Service Reference to your application and got the client proxy code. Your app works on your machine and you push it out. FAIL. NotFound. Crap. You forgot that your service reference had your local URI endpoint in there and when you moved it to staging and/or production it failed. You start cursing Microsoft and the Silverlight team and add to the threads in the forums or perhaps initiate a new wishlist item for the...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Cannot access my own web page

    - by enflam3
    I am developing, learning and experimenting with php,html,javascript,flash and so on. Having web hosting and all of the cpanel, phpmyadmin and other utilities. One day, while updating information, connection between my computer and website just went down. I found out that it is only from this computer, where I cannot access anything. I don't know what is the reason I cannot access website, however this is what I have checked so far: Everything else opens normally, having problem only with my page. Cannot access FTP,cpanel or any kind of information related to the domain and hosting ipconfigs detects IP, but shows request timed out (so its not browser related) Turned off Firewall,AV, Rebooted computer Cleared caches,temp,cookies,histry with CCleaner Checked connectivity with both (wired,wireless) networks ISP has dynamic IP that has been changed about 3 times since issue Checked host file I am out of ideas and understanding what could cause this kind of issue, however couple minutes ago, found out that everything works with proxy server (when adding IP and port to the browsers) Can someone point out what should I check or try to get rid of this problem?

    Read the article

< Previous Page | 137 138 139 140 141 142 143 144 145 146 147 148  | Next Page >